Volume 14 Supplement 19

## Highlights from the 2nd IEEE Symposium on Biological Data Visualization

# The looks of an odour - Visualising neural odour response patterns in real time

- Martin Strauch
^{1, 2}Email author, - Clemens Müthing
^{1}, - Marc P Broeg
^{1}, - Paul Szyszka
^{2}, - Daniel Münch
^{2}, - Thomas Laudes
^{2}, - Oliver Deussen
^{1}, - Cosmas Giovanni Galizia
^{2}and - Dorit Merhof
^{1}

**14(Suppl 19)**:S6

**DOI: **10.1186/1471-2105-14-S19-S6

© Strauch et al; licensee BioMed Central Ltd. 2013

**Published: **12 November 2013

## Abstract

### Background

Calcium imaging in insects reveals the neural response to odours, both at the receptor level on the antenna and in the antennal lobe, the first stage of olfactory information processing in the brain. Changes of intracellular calcium concentration in response to odour presentations can be observed by employing calcium-sensitive, fluorescent dyes. The response pattern across all recorded units is characteristic for the odour.

### Method

Previously, extraction of odour response patterns from calcium imaging movies was performed offline, after the experiment. We developed software to extract and to visualise odour response patterns in real time. An adaptive algorithm in combination with an implementation for the graphics processing unit enables fast processing of movie streams. Relying on correlations between pixels in the temporal domain, the calcium imaging movie can be segmented into regions that correspond to the neural units.

### Results

We applied our software to calcium imaging data recorded from the antennal lobe of the honeybee *Apis mellifera* and from the antenna of the fruit fly *Drosophila melanogaster*. Evaluation on reference data showed results comparable to those obtained by previous offline methods while computation time was significantly lower. Demonstrating practical applicability, we employed the software in a real-time experiment, performing segmentation of glomeruli - the functional units of the honeybee antennal lobe - and visualisation of glomerular activity patterns.

### Conclusions

Real-time visualisation of odour response patterns expands the experimental repertoire targeted at understanding information processing in the honeybee antennal lobe. In interactive experiments, glomeruli can be selected for manipulation based on their present or past activity, or based on their anatomical position. Apart from supporting neurobiology, the software allows for utilising the insect antenna as a chemosensor, e.g. to detect or to classify odours.

## Introduction

### Motivation

Odours take many shapes, and equipped with an insect brain and a neuroimaging device one can reveal these shapes, turning chemicals into patterns and images.

In the conference version of this paper [1], we have introduced an imaging system that can read out and process brain activity in real time, making the neural representations of odours accessible. The biological motivation is that access to ongoing brain activity is the basis for analysing storage and processing of information in the brain, observing, for example, the activity patterns in response to odour stimulation. In particular, transforming odours into patterns and images does not only benefit basic neuroscientific research: It also allows us to utilise a living organism with highly sensitive olfactory organs as a chemosensor, where the patterns and the distances between them contain information about odour identity and dissimilarity.

We consider two application scenarios for real-time visualisation of odours using insect brains. In insects, the first stage of odour perception is formed by the odour receptor neurons on the antenna. Here, we utilise calcium imaging to record from the antenna of the fruit fly *Drosophila melanogaster*. While such data provides only little information about signal processing in the brain, receptor neurons on the antenna are easy to access experimentally and they are excellent chemosensing devices. As such they are a promising alternative to artificial chemosensors, also referred to as electronic noses (see e.g. [2–5]), that find application in environmental monitoring, chemical industry or security.

The second stage of odour processing in the insect brain is the antennal lobe (AL), a dedicated olfactory center where odours are represented by activity patterns of neural units, the so-called glomeruli [6]. A network of interneurons connects the glomeruli, and unravelling the function of this network in processing odour information is the topic of ongoing research. The honeybee AL is an established model for studying odour learning and memory [7], and neuropharmacological tools [8, 9] have been developed to manipulate the network of interneurons. Here, the real-time aspect of odour visualisation is especially relevant as decisions can be based on prior activity, targeting e.g. glomeruli that have previously been part of a response pattern.

*Drosophila*antenna [11] are accessible through calcium imaging with fluorescent calcium reporters. Calcium imaging movies report fluorescence changes over time. Figure 1 gives an example for an imaging movie recorded from the honeybee AL, showing both the noisy raw images and processed, denoised versions of these images displayed in a false-colour scale.

The algorithm presented in this work computes a low-rank approximation based on few, selected (pixel) time series from the movie matrix. Exploiting the fact that there is noise and redundancy (time series from the same glomerulus are correlated) in the data, the movie matrix can be represented by another matrix of much lower rank. The rank-reduced version contains less noise, revealing the positions of the glomeruli and their signals [12]. As the glomerulus shapes become visible in Figure 1, such a representation may also be referred to as a segmentation, where the movie is segmented into regions with correlated activity over time, regions that correspond to the glomeruli.

Prior approaches to processing such imaging data involve manual feature selection [13, 14], or, if automated, they can only be performed offline [12, 15]. The real-time imaging system presented in this work allows for a wider range of applications, including closed-loop experiments, and it is what makes the readout from the antenna practical for chemosensing applications. Computation is performed in an adaptive manner, processing the movie stream incrementally as the images arrive from the camera. Fast processing is ensured by a GPGPU (General Purpose Computation on Graphics Processing Unit) implementation. The following sections provide further biological background and report on prior work. We then present algorithms and implementation details (Methods), followed by evaluation and demonstration of biological application cases (Results and discussion): Using the real-time imaging system, we visualise spontaneous activity and odour responses in the honeybee AL, and we provide a proof of concept for practical chemosensing with a biological sensor.

### Biological background

#### The olfactory system

The structure of olfactory systems is similar between insect species. As an example, we provide numbers for the honeybee *Apis mellifera*. On the antenna, approximately 60,000 odour receptor neurons interact physically with the odour molecules. These 60,000 neurons converge onto 160 glomeruli in the AL, where each glomerulus collects input from receptor neurons of one type. The glomeruli are bundles of synapses and appear as spherical structures with a diameter between 30 and 50 *µ* m. At the AL stage, each odour is represented by an activity pattern across the 160 glomeruli. Further upstream, this compact representation is widened again, as the projection neurons project from the glomeruli to approximately 160,000 Kenyon cells, a stage where odours are represented by sparser, high-dimensional patterns. [16, 17] While implementation details differ between species, the combinatorial coding of odours by activity patterns across glomeruli, in the insect AL or in the vertebrate olfactory bulb, is a common feature of olfactory systems and can also be found in humans [18].

#### Odour responses in the AL

Calcium imaging using calcium-sensitive fluorescent dyes grants us access to the odour response patterns in the AL of the honeybee *Apis mellifera* [10]. These odour response patterns reflect the response properties of the odour receptor neurons, as well as additional processing that takes place in the AL. Interneurons connecting the glomeruli perform further computations such as contrast enhancement [10].

For an example of odour response patterns in the honeybee AL, see Figure 1. The activity pattern of glomeruli (between ca. 20 and 40 of the glomeruli are visible in an imaging movie) fluctuates at a low amplitude when no odour is present. After stimulation with an odour (indicated by the black bar), glomeruli exhibit individual responses to the odour. As a result, the activity pattern across glomeruli changes in way that is characteristic for the odour. The same odour elicits a similar pattern in different bees [6].

There is evidence that not only the identity of a particular odour is encoded by the corresponding glomerular response pattern, but that also chemical [19] and perceptual [20] (dis)similarity are reflected by the (dis)similarity of response patterns, suggesting that response pattern space is a rather faithful representation of chemical space.

#### Odour responses on the antenna

Glomerular response patterns, as measured in the honeybee recordings from this work, are the output signal of the AL, i.e. they are the result of integrating all receptor neurons of one type and of further processing that occurs in the AL network of interneurons. While odour coding is improved after this processing [21], the results of this paper suggest that chemical identity and (dis)similarity can already be inferred from receptor neuron signals recorded on the antenna, the earliest stage in the olfactory processing pipeline where response patterns are easily accessible without dissecting the brain.

In this work, antenna data was recorded in the fruit fly *Drosophila melanogaster*. The genetic tools that are available for *Drosophila* make it possible to express a calcium reporter directly in the receptor neurons on the antenna. Instead of expressing the reporter in cells of one type, as has been done before [11], the approach pursued here is to measure signals from a large set of different receptor cells that all express the general olfactory co-receptor Orco that these cells bear in addition to a specific odour receptor. This allows us to measure broad odour response patterns across many receptors. The segmentation approach presented in this paper is then used to to extract individual response units from the imaging movie based on their differential responses to a series of 32 different odours.

### Related work

Computational approaches to analysing imaging data can be classified as being either synthetic or analytic. In a synthetic approach, similar to common procedures for analysing fMRI data, Stetter et al. [22] have set up non-linear functions that they fitted to the individual (pixel) time series of the imaging movie. These functions can account e.g. for dye bleaching over time and for different neural signal components. Rather than performing bottom-up synthetic reconstruction of the imaging movie, analytic approaches decompose (bottom-down) the movie into factors. These are matrix factorisation or decomposition methods that exist in many different flavours, e.g. the well-known Principal Component Analysis (PCA). In particular Independent Component Analysis (ICA) has found widespread application on imaging data [15, 23–26]. While ICA can be seen as a matrix decomposition method, the motivating paradigm for ICA is source separation. Under the assumption that there are underlying source signals that are statistically independent (and non-Gaussian), ICA algorithms (e.g. [27]) aim at recovering or separating these source signals on sample data where the sources appear in mixed form, e.g. neural signals mixed with measurement artifacts.

A recent convex analysis approach [12] (s.a. Methods), performs a factorisation of the movie matrix based on extremal column vectors from the boundary of the convex/conical hull of the data. Under the assumption that pure signal sources are present in the data, finding the extremal column vectors identifies these pure signal sources.

Traditionally, calcium imaging data from the insect AL has been processed by semi-automatic methods that perform e.g. image smoothing, but that still require human interaction to select regions of interest [13, 14]. From the methods listed above, those that require human interaction appear less suited for real-time processing on a movie stream. So far, no real-time implementations of the computational approaches exist, the software implementations from [12, 15] being only suited for offline data analysis.

## Methods

### Biological methods

#### Imaging the honeybee AL

For honeybees, it has been shown that projection neuron firing rate correlates with changes in intracellular calcium [28]. Staining with calcium-sensitive fluorescent dyes and excitation of the dyes with UV-light thus leads to a good proxy signal for brain activity [29].

*Apis mellifera*) was performed as described in [30]. Projection neurons in the l-APT and m-APT (lateral/medial antenno-protocerebral tract) were stained with Fura2-dextran (Invitrogen, Molecular Probes, Eugene, OR, USA), a calcium-sensitive, fluorescent dye. Activity of the projection neurons, that depart from the glomeruli in the AL, could be recorded using the experimental setup displayed in Figure 2. A fluorescence microscope (Axio Imager D.1, Zeiss, Göttingen, Germany) was equipped with a water immersion objective (20 ×, NA 0.95, Olympus, Tokyo, Japan). A light source (Polychrome V, TILL Photonics, Gräfelfing, Germany) provided excitation light at 340 and 380 nm, and fluorescence was recorded with a CCD camera (Andor Clara, Andor Technology PLC, Belfast, Northern Ireland). The input signal for data processing was computed as the ratio between consecutive images recorded at 340 and 380 nm, a standard procedure for Fura2-dextran [31].

#### Imaging the Drosophila antenna

##### Animals

Animals used for the experiments were female *Drosophila melanogaster* that were reared at 25 °C in a 12/12 light/dark cycle. Flies were of genotype w; P[Orco:Gal4]; P[UAS:GCaMP3]attP40, expressing the calcium reporter G-CaMP3 [32, 33] in all Orco (olfactory co-receptor) bearing cells (UAS-GCaMP3 flies were provided by Loren L. Looger, Howard Hughes Medical Institute, Janelia Farm Research Campus, Ashburn, Virginia, USA).

##### Odorant preparation

Odorants were purchased from Sigma-Aldrich in the highest purity available. Pure substances were diluted in 5 mL mineral oil (Sigma-Aldrich, Steinheim, Germany) to a concentration of 10^{-2} vol/vol. Odours were prepared in 20 mL headspace vials, covered with nitrogen and sealed with a Teflon septum (Axel Semrau, Germany). Odorants used were: 2-propylphenol (644-35-9), alpha-ionone (127-41-3), alpha-bisabolol (23089-26-1), trans-caryophyllene (87-44-5), (R)-carvone (6485-40-1), (S)-carvone (2244-16-8), beta-citronellol (106-22-9), 4-allyl-1,2-dimethoxybenzene (93-15-2), ethyl 3-hydroxyhexanoate (2305-25-1), ethyl (R)-3-hydroxybutanoate (24915-95-5), eugenol (97-53-0), E, E-farnesol (106-28-5), geraniol (106-24-1), heptyl acetate (112-06-1), hexyl acetate (142-92-7), hexyl butyrate (2639-63-6), isoamyl tiglate (41519-18-0), iso-eugenol (97-54-1), 4-isopropylbenzaldehyde (122-03-2), linalool (78-70-6), methyl 3-hydroxy hexanoate (21188-58-9), 4-methoxybenzaldehyde (123-11-5), methyl jasmonate (39924-52-2), (1R)-myrtenal (564-94-3), nonanal (124-19-6), nonanone (821-55-6), octyl acetate (112-14-1), phenylacetaldehyde (122-78-1), 4-hydroxy-3-methoxybenzaldehyde (121-33-5), gamma-propyl-gamma-butyrolactone (105-21-5), alpha-terpineol (10482-56-1) and alpha-thujone (546-80-5).

##### Stimulus application

A computer-controlled autosampler (PAL, CTC Switzerland) was used for automatic odour application. 2 mL of headspace was injected in two 1 mL portions at timepoints 6 s and 9 s with an injection speed of 1 mL/s into a continuous flow of purified air flowing at 60 mL/min. The stimulus was directed to the antenna of the animal via a Teflon tube (inner diameter 1 mm, length 38 cm).

The interstimulus interval was approximately 2 min. Solvent control and reference odorants (heptyl acetate and nonanone) were measured after every five stimuli (one block). The autosampler syringe was flushed with purified air for 30 s after each injection and washed with pentane (Merck, Darmstadt, Germany) automatically after each block of stimuli.

##### Calcium imaging

Calcium imaging was performed with a fluorescence microscope (BX51WI, Olympus, Tokyo, Japan) equipped with a 50x air lens (Olympus LM Plan FI 50x/0.5). A CCD camera (TILL Imago, TILL Photonics, Gräfelfing, Germany) was mounted on the microscope, recording with 4x4 pixel on-chip binning, resulting in 160x120 pixel sized images. For each stimulus, recordings of 20 s at a rate of 4 Hz were performed using TILL Vision (TILL Photonics, Gräfelfing, Germany).

A monochromator (Polychrome II, TILL Photonics, Gräfelfing, Germany) produced excitation light of 470 nm wavelength which was directed onto the antenna via a 500 nm low-pass filter and a 495 nm dichroic mirror. Emission light was filtered through a 505 nm high-pass emission filter.

Flies were mounted in custom-made holders, placed with their neck into a slit. The head was fixed to the holder with a drop of low-melting wax. A half electron microscopy grid was placed on top of the head, stabilising the antenna by touching the 2nd, but not the 3rd antennal segment.

### Matrix factorisation framework

*n*pixels into row vectors of length

*n*. The movie matrix

*A*

^{ m × n }has

*m*time points and

*n*pixels. The rows of the movie matrix,

*A*

_{(i)}, contain images or time points. The columns,

*A*

^{(j)}, contain pixels or time series.

*A*into a matrix

*T*

^{ m × k }of

*k*time series and a matrix

*S*

^{ k × n }of

*k*images, where $k\ll n,m.$ This provides a low-rank approximation

*A*

_{ k }to the original matrix

*A*:

In imaging movies, all pixels that report the signal of the same glomerulus are correlated with each other (apart from the noise), which causes redundancy in *A*. It is thus possible to construct a good approximation with small *k*, such that $\parallel A-TS\parallel $ is small.

The optimal rank-*k* approximation with respect to the aforementioned norm difference can be computed with Principal Component Analysis (PCA) [34]. However, the images in *S* computed by PCA are not sparse, with almost all pixels being different from zero [12]. The images in *S*, and the corresponding time series in *T*, can thus hardly be interpreted as the boundaries or the signal, respectively, of a particular neural unit. By definition, principal components need to be orthogonal to each other, which often prevents them from closely fitting the underlying source signals.

Ideally, as in the example from Figure 3, the images in *S* should be sparse, with only few pixels being different from zero. The *k* time series in *T* should be selected from *k* different glomeruli with the corresponding rows in the sparse *S* marking positions and boundaries of the glomeruli. We have shown in [12] that there is a method, the convex cone algorithm, that can achieve a factorisation with such favourable properties on imaging data.

### Convex cone algorithm

The movie matrix *A* can be described by basis time series in *T* that are combined by coefficients in the non-negative matrix *S*^{0+}. Residual noise is accounted for by *N*.

We assume that the columns *A*^{(j)}of the movie matrix contain either pure glomerulus signals or mixed signals, i.e. linear combinations (with non-negative coefficients) of the pure signals. At the fringes of a glomerulus, close to the neighbour glomeruli, such mixed signals can occur when a glomerulus signal is contaminated with additive light scatter from one or more neighbour glomeruli. Even if a glomerulus does not respond to an odour, light scatter can give the impression of a signal. In the middle of the glomeruli, that are rather large, circular objects, light scatter from the (distant) pixels of the neighbour glomeruli is less likely, and we assume that here the pixels contain pure signals.

For the matrix factorisation framework from Figure 3, we would like to select one pure signal from each glomerulus into *T*. Mixtures can then be modelled by *S*^{0+}. While *S*^{0+} can be computed easily given *A* and *T*, the challenging part is the selection of time series from the glomeruli into *T*.

Geometrically, the columns in *T* span a convex cone [35, 36] that contains a part of the data points in *A*. Data points that lie within the cone can be reconstructed exactly by linear combination (with non-negative coefficients) of the columns in *T* . Data points that lie outside of the cone can be approximated by projecting them to the boundary of the cone, where the approximation error depends on the distance to the boundary.

From convex analysis we know that the set of extreme vectors of *A* is the minimal generator of the convex cone that contains the entire *A* [35, 36]. With the extreme vectors we can span a volume that contains all data points of *A* and that thereby reduces the approximation error to zero. For imaging movies, the extreme columns vectors are also the columns with the pure signals from the middle of the glomeruli, whereas the mixed signal columns, that can be combined from the extreme, pure signal columns, lie within the cone.

Following this motivation, the convex cone algorithm [12] makes locally optimal choices for the next extreme column vector. With each new vector selected by the algorithm, the columns in *T* span a larger (≥) volume.

The convex cone algorithm starts with matrix *A*_{{1}} := *A*, selecting the column with index *p* that has the largest Euclidean norm: $\mathrm{arg}{\mathrm{max}}_{p}\left|\right|{A}_{\left\{1\right\}}^{(p)}\left|\right|$ This column becomes the first column of $T,{T}^{\left(1\right)}:={A}_{\left\{1\right\}}^{\left(p\right)}$. Then, a matching ${S}_{\left(1\right)}:={A}_{\left\{1\right\}}^{T}{T}^{\left(1\right)}$ is computed (for simplicity we omit the non-negativity constraint on *S*). The movie matrix is downdated as ${A}_{\left\{2\right\}}:={A}_{\left\{1\right\}}-{T}^{\left(1\right)}{S}_{\left(1\right)}$. In the new matrix *A*_{{2}} at iteration 2, the influence of the first column *T*^{(1)} is removed. We then select the column that is farthest away from the boundary of the cone, i.e. the column with the largest norm in *A*_{{2}}. This is an estimate for the next extreme column vector, and we fill this column into *T*^{(2)}.

We repeat the process until *c* columns are selected. In the following, we reserve *c* for the (user-specified) number of columns selected by the convex cone algorithm, and *k* for the number of principal components in the PCA step that is performed before the convex cone algorithm.

### Working on a movie stream

There are two motivations for performing PCA as a preprocessing prior to the convex cone algorithm. First, keeping only the top-*k* principal components reduces noise, which can make selection of extreme columns more robust. Second, we can utilise PCA to reduce computation time. For the real-time application, the movie matrix *A* grows by one row at each time point. The complexity of the convex cone algorithm is in the order $\mathcal{O}\left(mnc\right)$ if run once on the complete matrix *A*. For a growing movie matrix this would quickly accumulate a large overhead, the cost of performing the convex cone algorithm at each time point being $\mathcal{O}\left(1nc+2nc+\dots +mnc\right)$.

If we utilise PCA to keep, at all times, a compact summary matrix of constant size, we can remove the dependency on the growing time dimension. We propose to use an incremental PCA (IPCA) approach that computes the matrix *V*_{
k
} of the top-*k* principal components at each time point, where *V*_{
k
} is updated at low cost given the old version of *V*_{
k
} and the current image received from the movie stream. The convex cone algorithm is then no longer performed directly on *A*, but on *V*_{
k
}. As *V*_{
k
} is the minimiser of $\left|\right|A-{V}_{k}\left|\right|$, moderate values for *k* are sufficient in practice.

Several publications have treated IPCA algorithms [37–42]. Here, we rely on the CCIPCA algorithm by Weng et al. [39]. Several successful applications of CCIPCA can be demonstrated [43–45]. In these cases, CCIPCA was also used to incrementalise another algorithm by providing an updated version of matrix *V*_{
k
} at each time point. CCIPCA costs a constant $\mathcal{O}\left(nk\right)$ operations per update, which amounts to $\mathcal{O}\left(mnk\right)$ for processing the entire movie once.

Here, we outline the basic principle behind CCIPCA. The first principal component is approximated as the mean of the images received so far. The second principal component is approximated as the mean of the images from which the projection onto the first PC has been subtracted, etc. This approach allows for incremental updates, and it completely avoids the time-consuming construction of a large *n* × *n* covariance matrix, which would be required by standard PCA approaches that compute the eigenvectors of the covariance matrix. In the following, we briefly outline the CCIPCA iteration. For further details on CCIPCA, see [39]. A convergence proof is given in [46].

*A*

_{(i)}, at time point

*i*. The

*r*=1

*, ..., k*rows of the principal component matrix

*V*=

*V*

_{ k }are initialised with

*k*arbitrary, orthogonal vectors. Then,

*V*is upated at each time point using the current image

*A*

_{(i)}, where principal component

*V*

_{(r)}at time point

*i*is denoted ${V}_{\left(r\right)}^{\left\{i\right\}}$:

*A*

_{(i)}is downdated by subtracting the projection onto ${V}_{\left(r\right)}^{\left\{i\right\}}$:

After updating the *r* th principal component in this way, we can return to Equation (3) to update *V*_{(r+1)}(Algorithm 1).

**Algorithm 1:** *V*^{{i}}= Update_IPCA (*V* ^{{i−1}},*A*_{(i)}, *k* , *i*)

**for all**

*r*∈ [0,

*k*− 1]

**do**

**end for**

### Cone_updating: Visualisation in real time

Combining CCIPCA (Algorithm 1) and the convex cone algorithm leads to the algorithm at the core of the real-time imaging system: Cone_updating (Algorithm 2). Each image *A*(*i*) at time point *i* is first preprocessed by pixel-wise z-score normalisation: Subtract *µ*, the mean, and divide by *σ*, the standard deviation. Both, *µ* and *σ* of a pixel, can be updated as the movie stream proceeds.

After normalisation, the matrix *V* of the top-*k* principal components is updated with the current image: *V*^{{i}}:= Update_IPCA(*V*^{{i−1}}, *A*(*i*), *k*, *i*). Finally, the Convex_cone_algorithm (*V*^{{i}}, *c*) is applied to select *c* pixels (columns) from the current version of *V*.

As the movie matrix *A* grows, the incremental estimates for *µ*, *σ* and *V* improve. As a consequence, the *c* columns selected by the convex cone algorithm are better estimates of the extreme column vectors of *A*, the vectors that contain the pure glomerulus signals. In the matrix factorisation framework, these are are the columns for matrix *T*, and the corresponding *S* indicates glomerulus position (Figure 3).

Visualisations of brain activity, such as in Figure 1, can be achieved by low-rank approximation, using matrices T and S: *A*_{
k
} = *TS*. At time point *i* we do not yet know the final *T* and *S*, and therefore we obtain the approximation ${\widehat{A}}_{(i)}\phantom{\rule{0.1em}{0ex}}as{\widehat{A}}_{(i)}:={A}_{(i)}{S}^{\left\{i\right\}}{S}^{\left\{i\right\}}$, where ${S}^{\left\{i\right\}}$ is the current version of *S*.

For offline data visualisation, the colour scale can be adjusted to the maximum and minimum value of *A*. For real-time display, using one colour scale for the entire movie, maximum and minimum have to be updated incrementally. To avoid level changes, e.g. by long-term photobleaching of the calcium dye, data was high-pass filtered (0:025 Hz) before display in a false-colour scale (as in Figure 1).

**Algorithm 2** : *S* = Cone_updating (*A*^{(m × n)}, *c*, *k*)

Initialise ${V}^{\left\{1\right\}}$

**for all** *i* ∈ [0, *m* - 1] **do**

*A*_{(i)}:= z_score_normalise (*A*_{(i)})

**if** *i* > 1 **then**

*V*^{{i}}:= Update_IPCA(*V* ^{{i-1}},*A*_{(i)}, *k*, *i*)

*S*^{{i}}:= Convex_cone_algorithm (*V*^{{i}}, *c*)

${\widehat{A}}_{\left(i\right)}:={A}_{\left(i\right)}{S}^{\left\{i\right\}}{S}^{\left\{i\right\}}//$ low-rank approximation to image *A*_{(i)}

**end if**

**end for**

### Implementations

We consider three implementations of the convex cone algorithm. Two implementations were written in Java, Java_offline, the reference implementation from [12], and Java_online. Java_offline performs exact offline PCA, followed by the convex cone algorithm, whereas Java_online (implementation of Algorithm 1) uses incremental PCA instead. Both Java implementations were performed in KNIME [47] (http://www.knime.org), a data pipelining environment.

Finally, we implemented the incremental online variant (Algorithm 1) using GPGPU: GPGPU_online. Z-score normalisation and the time-consuming PCA were implemented for the GPU with the NVIDIA CUDA [48] Basic Linear Algebra Subroutines (cuBLAS) (http://developer.nvidia.com/cublas) and the CUDA Linear Algebra library (CULA) (http://www.culatools.com/). The actual convex cone algorithm was run on the CPU.

TILL Photonics Live Acquisition (LA) Software 2.0 [49] was employed to control experimental hardware and excitation light intensity. GPGPU_online accessed the movie stream directly from the camera using a custom-built software interface kindly provided by TILL Photonics.

## Results and discussion

This sections starts with a technical evaluation of the proposed algorithm and a comparison of different implementations. We then demonstrate practical applicability in an experiment with honeybees and show how the techniques developed in this work can be utilised to turn the insect antenna into a living chemosensor. We conclude with a discussion regarding the impact that real-time processing of neural activity can have.

### Performance measures

#### Computing time

The motivations for adapting the matrix factorisation framework to the datastream domain were the ability to perform incremental updates upon arrival of new data, and of course the ability to process data with minimal time delay. For evaluation, we performed computation time measurements on a reference dataset. Measurements were carried out using an Intel Core i7 950 (3.07 GHz) CPU and a NVIDIA.

GeForce GTX 285 (648 MHz, 1024 MB) GPU. The Java_offline and Java_online implementations were run in KNIME (http://www.knime.org) workflows, and, for comparability with the C-implementation GPGPU_online, time measurements do only include the actual computation time and not the time for data transfer between nodes in the KNIME workflow.

Computation time measurements were performed on the 11 imaging movies of the reference dataset.

Implementation | (ms/frame) | overall time consumption (min) |
---|---|---|

Java_offline | 134 | 68.15 |

Java_online | 65 | 39.26 |

GPGPU_online | 23 | 18.31 |

With the fastest implementation, GPGPU_online, a single image from the movie can be processed in 23 ms (Table 1), which is sufficient for calcium imaging in honeybees and *Drosophila* with typical recording frequencies below 20 Hz.

#### Approximation quality

*S*, the images that show the positions and boundaries of the glomeruli (see Figure 3). Such glomerular maps reveal the anatomy of the AL and can be matched between bees [12]. Using the reference results from [12], we compared glomerulus maps computed by Java_offline and GPGPU_online on the same movie. Parameters were

*k*= 50 principal components,

*c*= 50 (convex cone algorithm). Figure 4a shows glomerulus maps constructed by the two implementations. Clearly, both implementations reveal the same anatomy, but there is no perfect correspondence between the maps.

Apart from visual inspection of glomerulus maps, where cluster size has a strong impact, we also analysed how robust signal (column) selection was against incremental approximation. The convex cone algorithm selects extreme column vectors into matrix *T* that correspond to the pure signal sources. Matrix *S* is computed given *A* and *T* , and the rows of *S* reflect the distribution of similarity with the corresponding signal sources in *T* . This gives rise to the clusters of similar pixels in *S* and on the glomerulus map. Preprocessing, such as incremental z-score normalisation and incremental PCA, and a postprocessing step employed in [12] to remove the residual noise *N* (Equation 2) all have an impact on signal similarity and thus influence cluster size. This affects especially clusters that correspond to e.g. areas of similarly strong background staining, illumination artifacts etc. that do not have such clearly distinct signals as the glomeruli.

To evaluate signal selection independent of cluster size, we visualised the positions of the columns selected by the offline reference implementation Java_offline, along with the positions of the columns selected by GPGPU_online. For the reference implementation, we included only glomerular signals, i.e. those that could be identified by matching glomerulus maps to the anatomical honeybee AL atlas [50]. Figure 4b shows that the positions of signals selected by GPGPU_online (black circles) are in good correspondence with the glomerulus "targets" provided by Java_offline (red triangles). We conclude that selection of relevant signals, i.e. glomerular signals, is robust against incremental approximation in the online setting.

### Documentation of a real-time experiment in the honeybee AL

Additional file 1: Video documentation, part 1. Experimental setup for honeybee brain imaging. (MP4 7 MB)

Additional file 2: Video documentation, part 2. Screen capture from a honeybee brain imaging experiment. (MP4 5 MB)

*c*basis signals were still influenced by the initialisation of the (not yet converged) incremental PCA (row of points in the left upper corner), they quickly moved to the positions of the glomeruli as more information arrived from the stream. Already after 400-600 time points, the map was almost complete. If needed, such an incremental computation of the glomerulus map could also be used to adapt to changes: For example, shifts between images caused by animal movement could be corrected for by giving higher weights to more recent time points.

### The insect antenna as a chemosensor

#### Functional segmentation of the antenna

While recording glomerular activity in the honeybee AL can help to answer neurobiological questions, the experimental procedure is technically demanding and time-consuming as it requires to dissect the brain and to fill in a calcium-dye into the projection neurons the day before the experiment. Working with the model organism *Drosophila melanogaster*, we could omit the manual staining step, genetically expressing the calcium-sensitive dye G-CaMP [32, 51] in the receptor neurons that are accessible without dissecting the brain (see Methods).

This experimental design allows us to employ a *Drosophila* antenna as an easy-to-handle chemosensor. Similar to the situation in the honeybee AL, stimulating the antenna with a series of odour presentations elicits differential responses in the receptors that can then be distinguished based on their activity over time. We can thus utilise the real-time software to construct a map of the *Drosophila* antenna and to observe odour responses.

*c*= 100 response units. The antenna map is shown in 8a, and Figure 8b contains a single image from the calcium imaging movie.

In *Drosophila*, olfactory receptor neurons are anisotropically spatially scattered over the surface of the antenna, so that odour response patterns form spatio-temporal maps. However, sources could be overlapping, creating a situation that is more complex than with the image of well segregated olfactory glomeruli in the antennal lobe.

#### Data analysis

*T*. For each odour stimulation, we computed a feature vector, where the feature was the maximum value of the response unit after presentation of the respective odour (and before the start of the next odour measurement): Figure 9a.

Clustering these feature vectors (Figure 9b) shows that feature vectors for repeated applications of the same odour, e.g. nonanone, cluster together. The odourless control measurements (mineral oil, air, N2) appear clearly separated from the odourous substances, and chemically similar odours end up in the same cluster (ethyl- and methyl-3-hydroxy-methanoate). This serves as a proof of concept, demonstrating that the real-time imaging system can, in principle, both recognise known odours and estimate the identity of unknown odours by their similarity to reference odours.

While distances between odour molecules are in part well reflected by response pattern distances, this is not always the case. For example, iso-eugenol does not fit into the heptyl acetate cluster, and 2-propylphenol lacks clear responses and therefore ends up in the (odourless) oil/air cluster. Further experiments are needed to evaluate whether representation of chemical identity can be optimised by recording more or different response units, e.g. in a different focal plane.

It also needs to be tested whether the observed odour responses are stereotypical across many individuals. As a first approach, we replicated the experiment from Figure 9, finding a high correlation (Pearson correlation 0.86, *p* =0.001, Mantel test for correlation of distance matrices) between the odour × odour Euclidean distance matrices (based on the feature vectors) from both experiments, indicating that the relative dissimilarity of odours could be conserved between individuals.

#### How can the system be applied?

Artificial chemosenors, so-called electronic noses [2–5] are important tools for environmental monitoring, healthcare or security applications. They do, however, not yet reach the efficiency and sensitivity of biological olfactory systems. The real-time software can extract features from calcium imaging recordings, directly accessing the *Drosophila* antenna as a biological chemosensor. Such feature vectors (Figure 9a) can be used to visualise molecular identity, or they can be subject to further processing, e.g. by classifiers, aiming to determine the identity of an unknown chemical substance. There are two points to making the biological chemosensor practical: 1) Working with non-invasive biological techniques that allow for easy handling of the flies, 2) Software that can process the continuous stream of odour plumes encountered in a real-world application.

### Impact of real-time processing

Going beyond the specific example of the chemosensor application, real-time processing has a wider range of applicability that involves any kind of interactive experimentation. This belongs to future work that is made possible with the real-time technology.

#### Motivating examples

It is increasingly clear that perception is influenced by both the stimulus and the prior state of the brain. For example, brain oscillations during the pre-stimulus interval influence how a human subject perceives an auditory stimulus in an experimental setup targeted at multimodal sensory integration [52]. Sensory experience without external stimulation, stemming only from the current state of the brain, is known from medical phenomena such as tinnitus [53].

For honeybees, there is first evidence in the direction that spontaneous background activity of the glomeruli in the AL carries information about odours that have been encountered recently: Glomerular activity patterns similar to a particular odour response pattern reverberate minutes after the actual response has been elicited by odour stimulation [54].

Considering the growing interest in ongoing brain activity, it is increasingly important to develop experimental strategies that allow stimulus presentations to be conditional on ongoing brain activity states. With the real-time methods presented in this work, glomeruli can be targeted because of their responses to odours or because they are part of reverberating patterns in spontaneous background activity.

Real-time processing is necessary to answer fundamental questions regarding the role of ongoing brain activity: Is it a side-effect that simply occurs as a consequence of neuron and network properties? Are patterns in spontaneous activity actually read out for further processing in the brain? In conditioning experiments [7], bees learn to associate an odour with a sugar reward. Can rewarding a pattern in spontaneous activity have the same effect as rewarding the actual odour?

#### Added value by real-time processing

From a biological perspective, the added value provided by the real-time software is that brain activity can be interpreted based on processed information. Only milliseconds after the activity occurs, we can regard not only raw pixel values, but anatomically distinct and identifiable units, the olfactory glomeruli in our case.

While analysis of neural data is often is performed pixel-wise (or voxel-wise), the brain encodes odours in patterns across glomeruli. Being able to work on a glomerulus level allows us to match the odour response patterns we observe with known response patterns from a database, which can reveal the chemical identity of the stimulus molecule. For spontaneous background activity, we can analyse the distribution of glomerular patterns that informs us about the state the antennal lobe network is in, i.e. the prior state that is relevant for how the stimulus will be perceived.

#### How fast is fast enough?

Closed loop experiments, where measured brain activity controls experimental settings, require that data processing is faster than recording speed. In calcium imaging experiments, images are often recorded at frequencies of 20 Hz or slower. Thus, any processing of 50 ms/frame or faster is appropriate. Recordings with voltage-sensitive dyes, for example, are generally useful at 50 Hz or faster: The fastest neuronal processes, the action potentials, have a duration of 1-3 ms, so recordings at 1000 Hz would be ideal. The current speed of 23 ms/frame (Table 1) is already getting close to the 50 Hz value, but it is still far from the ideal 1000 Hz.

For many experiments, fast processing is a requirement, e.g. if we wish to follow and react to fluctuations in spontaneous activity. For the chemosensor task, the advantage lies in the fact that we can directly query a biological chemosensor instead of waiting for results from post-hoc data analysis. Fast processing reduces the delay of the chemical analysis and allows for high-throughput assays.

## Conclusions

In the brain, odours are represented as activity patterns across many neurons. Calcium imaging is a technique that lends itself to extracting such activity patterns, as it allows to record many units simultaneously. So far, software for calcium imaging data has focussed on *offline* data processing [12–15]. The algorithms and software presented in this work process calcium imaging movies *online*, making the neural representations of odours accessible directly when they occur.

Algorithmically, we rely on a matrix factorisation that is updated with every new image that arrives from the movie stream. A low-rank approximation to the movie matrix serves as a compact representation of the calcium imaging movie, discarding noise and highlighting neural signals. This serves as the basis for further visualisations, such as functional maps of the glomeruli in the AL: Glomerulus borders are not defined by anatomy, but by function, i.e. activity (in response to odours) of pixels over time. This eliminates the need for registration of imaging data to anatomical stainings.

Such maps and the visualisation obtained by low-rank approximation reveal the "looks of an odour", the initial odour response pattern on the antenna, or, after data integration and processing has taken place, the glomerular response pattern in the AL.

Both odour representations have applications that profit from real-time processing. The role of the AL network in shaping the odour response patterns can now be investigated using closed-loop experiments, where prior system states influence current experimental parameters. Staining an array of receptor neurons with a single genetic construct, accompanied by online processing, provides easy access to odour response patterns, making real-time chemosensing with a biological sensor practical.

Visualising the neural representation of odours serves also to map perceptional spaces. Distances between odour response patterns are an estimate for perceptional (dis)similarity between odours [20] throughout different stages of odour information processing in the same individual, and also between individuals and even species, leading to species-specific odour perception spaces.

For such and further applications, the algorithmic and visualisation framework developed here enables fully automatic processing of odour response data without the need for human interaction to define e.g. regions of interest.

## Availability

Source code is available in Additional file 3.

## Declarations

### Acknowledgements

We are grateful to TILL Photonics GmbH (Munich, Germany) for providing a software interface for their imaging system. We would like to thank Julia Rein for the honeybee reference data, and Jacob Stierle for assistance with handling and preparing bees.

**Declarations**

This publication was funded by the IEEE Symposium on Biological Data Visualization (BioVis) as a supplement of highlights.

The articles in this supplement have undergone the journal's standard peer review process for supplements. The Supplement Editors declare that they have no competing interests.

This article has been published as part of *BMC Bioinformatics* Volume 14 Supplement 19, 2013: Highlights from the 2nd IEEE Symposium on Biological Data Visualization. The full contents of the supplement are available online at http://www.biomedcentral.com/bmcbioinformatics/supplements/14/S19.

## Authors’ Affiliations

## References

- Strauch M, Broeg MP, Müthing C, Szyszka P, Deussen O, Galizia CG, Merhof D: INCIDE the Brain of a Bee - Visualising Honeybee Brain Activity in Real Time by Semantic Segmentation. Proceedings of BioVis, Oct 14-15 2012, Seattle, USA. 2012, IEEE, 17-24.Google Scholar
- D'Amico A, Natale CD, Paolesse R, Macagnano A, Martinelli E, Pennazza G, Santonico M, Bernabei M, Roscioni C, Galluccio G, Bono R, Agro EF, Rullo S: Olfactory systems for medical applications. Sensors and Actuators B: Chemical. 2008, 130: 458-465. 10.1016/j.snb.2007.09.044.View ArticleGoogle Scholar
- Di Natale C, Martinelli E, Paolesse R, D'Amico A, Filippini D, Lundström I: An Experimental Biomimetic Platform for Artificial Olfaction. PLoS ONE. 2008, 3 (9): e3139-10.1371/journal.pone.0003139.PubMed CentralView ArticlePubMedGoogle Scholar
- Haddad R, Medhanie A, Roth Y, Harel D, Sobel N: Predicting Odor Pleasantness with an Electronic Nose. PLoS Comput Biol. 2010, 6 (4): e1000740-10.1371/journal.pcbi.1000740.PubMed CentralView ArticlePubMedGoogle Scholar
- Eifler J, Martinelli E, Santonico M, Capuano R, Schild D, Di Natale C: Differential Detection of Potentially Hazardous Fusarium Species in Wheat Grains by an Electronic Nose. PLoS ONE. 2011, 6 (6): e21026-10.1371/journal.pone.0021026.PubMed CentralView ArticlePubMedGoogle Scholar
- Galizia CG, Sachse S, Rappert A, Menzel R: The glomerular code for odor representation is species specific in the honeybee Apis mellifera. Nat Neurosci. 1999, 2 (5): 473-478. 10.1038/8144.View ArticlePubMedGoogle Scholar
- Rath L, Giovanni Galizia C, Szyszka P: Multiple memory traces after associative learning in the honeybee antennal lobe. European Journal of Neuroscience. 2011, 34 (2): 352-360. 10.1111/j.1460-9568.2011.07753.x.View ArticlePubMedGoogle Scholar
- Girardin C, Galizia C: The "Where" and "Who" in Brain Science: Probing Brain Networks with Local Perturbations. Cognitive Computation. 2012, 4: 63-70. 10.1007/s12559-011-9122-3.View ArticleGoogle Scholar
- Girardin CC, Kreissl S, Galizia CG: Inhibitory connections in the honeybee antennal lobe are spatially patchy. J Neurophysiol. 2013, 109 (2): 332-343. 10.1152/jn.01085.2011.View ArticlePubMedGoogle Scholar
- Sachse S, Galizia C: Role of inhibition for temporal and spatial odor representation in olfactory output neurons: a calcium imaging study. J Neurophysiol. 2002, 87 (2): 1106-1117.PubMedGoogle Scholar
- Pelz D, Roeske T, Syed Z, Bruyne Md, Galizia CG: The molecular receptive range of an olfactory receptor in vivo (Drosophila melanogaster Or22a). J Neurobiol. 2006, 66 (14): 1544-1563. 10.1002/neu.20333.View ArticlePubMedGoogle Scholar
- Strauch M, Rein J, Galizia CG: Signal extraction from movies of honeybee brain activity by convex analysis. Proceedings of ICCABS, Feb 23-25 2012. 2012, Las Vegas, USA, IEEE, 1-6.Google Scholar
- Ditzen M: Odor concentration and identity coding in the antennal lobe of the honeybee Apis mellifera. PhD thesis. 2005, Department of Biology, Freie Universit¨at Berlin, [http://www.diss.fu-berlin.de/2005/211/indexe.html]Google Scholar
- Dupuy F, Casasa J, Bagneresa AG, Lazzaria CR: OpenFluo: A free open-source software for optophysiological data analyses. J Neurosci Methods. 2009, 183 (2): 195-201. 10.1016/j.jneumeth.2009.06.031.View ArticlePubMedGoogle Scholar
- Mukamel EA, Nimmerjahn A, Schnitzer MJ: Automated Analysis of Cellular Signals from Large-Scale Calcium Imaging Data. Neuron. 2009, 63 (6): 747-760. 10.1016/j.neuron.2009.08.009.PubMed CentralView ArticlePubMedGoogle Scholar
- Galizia CG, Menzel R: The role of glomeruli in the neural representation of odors: results from optical recording studies. J Insect Phys. 2001, 47: 115-129. 10.1016/S0022-1910(00)00106-2.View ArticleGoogle Scholar
- Wilson RI, Mainen ZF: Early Events in Olfactory Processing. Annual Review of Neuroscience. 2006, 29: 163-201. 10.1146/annurev.neuro.29.051605.112950.View ArticlePubMedGoogle Scholar
- Malnic B, Hirono J, Sato T, Buck LB: Combinatorial Receptor Codes for Odors. Cell. 1999, 96 (5): 713-723. 10.1016/S0092-8674(00)80581-4.View ArticlePubMedGoogle Scholar
- Haddad R, Khan R, Takahashi YK, Mori K, Harel D, Sobel N: A metric for odorant comparison. Nat Methods. 2008, 5: 425-429. 10.1038/nmeth.1197.View ArticlePubMedGoogle Scholar
- Guerrieri F, Schubert M, Sandoz JC, Giurfa M: Perceptual and neural olfactory similarity in honeybees. PLoS Biol. 2005, 3 (4): e60-10.1371/journal.pbio.0030060.PubMed CentralView ArticlePubMedGoogle Scholar
- Sachse S, Galizia CG: The coding of odour-intensity in the honeybee antennal lobe: local computation optimizes odour representation. Eur J Neurosci. 2003, 18 (8): 2119-2132. 10.1046/j.1460-9568.2003.02931.x.View ArticlePubMedGoogle Scholar
- Stetter M, Greve H, Galizia CG, Obermayer K: Analysis of calcium imaging signals from the honeybee brain by nonlinear models. NeuroImage. 2001, 13: 119-128.View ArticlePubMedGoogle Scholar
- Brown GD, Yamada S, Sejnowski TJ: Independent component analysis at the neural cocktail party. Trends in Neurosciences. 2001, 24: 54-63. 10.1016/S0166-2236(00)01683-0.View ArticlePubMedGoogle Scholar
- Reidl J, Starke J, Omer D, Grinvald A, Spors H: Independent component analysis of high-resolution imaging data identifies distinct functional domains. NeuroImage. 2007, 34: 94-108. 10.1016/j.neuroimage.2006.08.031.View ArticlePubMedGoogle Scholar
- Strauch M, Galizia CG: Registration to a Neuroanatomical Reference Atlas-Identifying Glomeruli in Optical Recordings of the Honeybee Brain. Proceedings of GCB, Sep 9-12 2008, Dresden, Germany. 2008, GI, 85-95. Lectures Notes in InformaticsGoogle Scholar
- Strauch M, Galizia CG: Fast PCA for Processing Calcium-Imaging Data from the Brain of Drosophila melanogaster. Proceedings of DTMBIO, Oct 24th 2011, Glasgow, Scotland. 2011, ACM, 3-10.Google Scholar
- Hyvärinen A: Fast and Robust Fixed-Point Algorithms for Independent Component Analysis. IEEE Transactions on Neural Networks. 1999, 10 (3): 626-634. 10.1109/72.761722.View ArticlePubMedGoogle Scholar
- Galizia CG, Kimmerle B: Physiological and morphological characterization of honeybee olfactory neurons combining electrophysiology, calcium imaging and confocal microscopy. J Comp Physiol A. 2004, 190: 21-38. 10.1007/s00359-003-0469-0.View ArticleGoogle Scholar
- Grienberger C, Konnerth A: Imaging Calcium in Neurons. Neuron. 2012, 73 (5): 862-885. 10.1016/j.neuron.2012.02.011.View ArticlePubMedGoogle Scholar
- Szyszka P, Demmler C, Oemisch M, Sommer L, Biergans S, Birnbach B, Silbering AF, Galizia CG: Mind the Gap: Olfactory Trace Conditioning in Honeybees. J Neurosci. 2011, 31 (20): 7229-7239. 10.1523/JNEUROSCI.6668-10.2011.View ArticlePubMedGoogle Scholar
- O'Connor N, Silver RB: Ratio Imaging: Practical Considerations for Measuring Intracellular Ca
^{2+}and pH in Living Cells. Digital Microscopy, 3rd Edition, Volume 81 of Methods in Cell Biology. Edited by: Sluder G, Wolf DE. 2007, Academic Press, 415-433.View ArticleGoogle Scholar - Nakai J, Ohkura M, Imoto K: A high signal-to-noise Ca2+ probe composed of a single green fluorescent protein. Nat Biotech. 2001, 19 (2): 137-141. 10.1038/84397.View ArticleGoogle Scholar
- Tian L, Hires SA, Mao T, Huber D, Chiappe ME, Chalasani SH, Petreanu L, Akerboom J, McKinney SA, Schreiter ER, Bargmann CI, Jayaraman V, Svoboda K, Looger LL: Imaging neural activity in worms, flies and mice with improved GCaMP calcium indicators. Nat Meth. 2009, 6 (12): 875-881. 10.1038/nmeth.1398.View ArticleGoogle Scholar
- Jolliffe IT: Principal Component Analysis. 2002, New York/Berlin/Heidelberg: Springer, [ISBN: 0387954422], 2Google Scholar
- Rockafellar RT: Convex Analysis. 1970, Princeton, NJ: Princeton University Press, [ISBN: 0691080690]View ArticleGoogle Scholar
- Dattoro J: Convex Optimisation & Euclidean Distance Geometry. 2011, Meboo Publishing, Palo Alto, CA, [ISBN: 0976401304]Google Scholar
- Skocaj D, Leonardis A: Weighted and Robust Incremental Method for Subspace Learning. Proceedings of ICCV, Oct 14-17 2003, Nice, France. 2003, IEEE, 1494-1501.Google Scholar
- Skocaj D, Leonardis A, Bischof H: Weighted and robust learning of subspace representations. Pattern Recognition. 2007, 40 (5): 1556-1569. 10.1016/j.patcog.2006.09.019.View ArticleGoogle Scholar
- Weng J, Zhang Y, Hwang W: Candid Covariance-free Incremental Principal Component Analysis. IEEE Trans Pattern Analysis and Machine Intelligence. 2003, 25: 1034-1040. 10.1109/TPAMI.2003.1217609.View ArticleGoogle Scholar
- Yan S, Tang X: Largest-eigenvalue-theory for incremental principal component analysis. International Conference on Image Processing (ICIP), Sep 11-14 2005, Genoa, Italy. 2005, IEEE, 1181-1184.Google Scholar
- Zhao H, Yuen PC, Kwok JT: A novel incremental principal component analysis and its application for face recognition. IEEE Transactions on Systems, Man, and Cybernetics B. 2006, 36: 873-886.View ArticleGoogle Scholar
- Huang D, Yi Z, Pu X: A New Incremental PCA Algorithm With Application to Visual Learning and Recognition. Neural Processing Letters. 2009, 30 (3): 171-185. 10.1007/s11063-009-9117-1.View ArticleGoogle Scholar
- Yan J, Zhang B, Yan S, Yang Q, Li H, Chen Z, Xi W, Fan W, Ma WY, Cheng Q: IMMC: incremental maximum margin criterion. Proceedings of SIGKDD, Aug 22-25 2004, Seattle, USA. 2004, ACM, 725-730.Google Scholar
- Dagher I, Nachar R: Face Recognition Using IPCA-ICA Algorithm. IEEE Trans Pattern Anal Mach Intell. 2006, 28 (6): 996-1000.View ArticlePubMedGoogle Scholar
- Kompella VR, Luciw MD, Schmidhuber J: Incremental Slow Feature Analysis. Proceedings of IJCAI, Jul 16-22 2011, Barcelona, Spain. 2011, AAAI Press, 1354-1359.Google Scholar
- Zhang Y, Weng J: Convergence analysis of complementary candid incremental principal component analysis. Tech rep, Comput Sci Eng, Michigan State Univ. 2001Google Scholar
- Berthold M, Cebron N, Dill F, Gabriel T, Kötter T, Meinl T, Ohl P, Sieb C, Thiel K, Wiswedel B: KNIME: The Konstanz Information Miner. Data Analysis, Machine Learning and Applications. Edited by: Preisach C, Burkhardt H, Schmidt-Thieme L, Decker R. 2008, Springer Berlin Heidelberg, 319-326. Studies in Classification, Data Analysis, and Knowledge OrganizationView ArticleGoogle Scholar
- Corporation N: NVIDIA CUDA Compute Unified Device Architecture Programming Guide. 2007, NVIDIA CorporationGoogle Scholar
- TILL Photonics: Live Acquisition Software. Website. 2011, [http://www.till-photonics.com/Products/la.php]Google Scholar
- Galizia CG, McIlwrath SL, Menzel R: A digital three-dimensional atlas of the honeybee antennal lobe based on optical sections acquired by confocal microscopy. Cell Tissue Res. 1999, 295 (3): 383-394. 10.1007/s004410051245.View ArticlePubMedGoogle Scholar
- Wang JW, Wong AM, Flores J, Vosshall LB, Axel R: Two-Photon Calcium Imaging Reveals an Odor-Evoked Map of Activity in the Fly Brain. Cell. 2003, 112 (2): 271-282. 10.1016/S0092-8674(03)00004-7.View ArticlePubMedGoogle Scholar
- Keil J, Müller N, Ihssen N, Weisz N: On the Variability of the McGurk Effect: Audiovisual Integration Depends on Prestimulus Brain States. Cerebral Cortex. 2012, 22: 221-231. 10.1093/cercor/bhr125.View ArticlePubMedGoogle Scholar
- Weisz N, Dohrmann K, Elbert T: The relevance of spontaneous activity for the coding of the tinnitus sensation. Tinnitus: Pathophysiology and Treatment, Volume 166 of Progress in Brain Research. Edited by: Langguth B, Hajak G, Kleinjung T, Cacace A, Møller A. 2007, Elsevier, 61-70.View ArticleGoogle Scholar
- Galàn RF, Weidert M, Menzel R, Herz AVM, Galizia CG: Sensory memory for odors is encoded in spontaneous correlated activity between olfactory glomeruli. Neural Comput. 2006, 18: 10-25. 10.1162/089976606774841558.View ArticlePubMedGoogle Scholar
- Brandt R, Rohlfing T, Rybak J, Krofczik S, Maye A, Westerhoff M, Hege HC, Menzel R: Three-dimensional average-shape atlas of the honeybee brain and its applications. J Comp Neurol. 2005, 492: 1-19. 10.1002/cne.20644.View ArticlePubMedGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.