- Research
- Open access
- Published:
Automatic plankton image classification combining multiple view features via multiple kernel learning
BMC Bioinformatics volume 18, Article number: 570 (2017)
Abstract
Background
Plankton, including phytoplankton and zooplankton, are the main source of food for organisms in the ocean and form the base of marine food chain. As the fundamental components of marine ecosystems, plankton is very sensitive to environment changes, and the study of plankton abundance and distribution is crucial, in order to understand environment changes and protect marine ecosystems. This study was carried out to develop an extensive applicable plankton classification system with high accuracy for the increasing number of various imaging devices. Literature shows that most plankton image classification systems were limited to only one specific imaging device and a relatively narrow taxonomic scope. The real practical system for automatic plankton classification is even non-existent and this study is partly to fill this gap.
Results
Inspired by the analysis of literature and development of technology, we focused on the requirements of practical application and proposed an automatic system for plankton image classification combining multiple view features via multiple kernel learning (MKL). For one thing, in order to describe the biomorphic characteristics of plankton more completely and comprehensively, we combined general features with robust features, especially by adding features like Inner-Distance Shape Context for morphological representation. For another, we divided all the features into different types from multiple views and feed them to multiple classifiers instead of only one by combining different kernel matrices computed from different types of features optimally via multiple kernel learning. Moreover, we also applied feature selection method to choose the optimal feature subsets from redundant features for satisfying different datasets from different imaging devices. We implemented our proposed classification system on three different datasets across more than 20 categories from phytoplankton to zooplankton. The experimental results validated that our system outperforms state-of-the-art plankton image classification systems in terms of accuracy and robustness.
Conclusions
This study demonstrated automatic plankton image classification system combining multiple view features using multiple kernel learning. The results indicated that multiple view features combined by NLMKL using three kernel functions (linear, polynomial and Gaussian kernel functions) can describe and use information of features better so that achieve a higher classification accuracy.
Background
Plankton, including phytoplankton and zooplankton, are the main source of food for organisms in the ocean and form the base of marine food chain. As the fundamental components of marine ecosystems, plankton is very sensitive to environment changes, and its abundance plays an important role on the ocean ecological balance. Therefore, the study of plankton abundance and distribution is crucial, in order to understand environment changes and protect marine ecosystems.
In the early days, researchers investigated the distribution and abundance of plankton with traditional techniques, such as Niskin bottles, pumps and towed nets, to collect the samples. And then, the classification and counting were done manually by experts. These traditional methods for the study of plankton are so laborious and time consuming that hindered the understanding process of plankton.
To improve the efficiency, many imaging devices, including in situ and in the lab, have been developed for collecting plankton images, such as Video Plankton Recorder (VPR) [1], Underwater Video Profiler (UVP) [2], Shadowed Image Particle Profiling Evaluation Recorder (SIPPER) [3], Zooplankton Visualization System (ZOOVIS) [4], Scripps Plankton Camera (SPC) [5], Imaging FlowCytobot (IFCB) [6], In Situ Ichthyoplankton Imaging System (ISIIS) [7], ZooScan [8], and so on. These imaging devices are able to generate an enormous amount of plankton images within a short time. However, if these collected images are manually classified and counted, there will be a daunting task. Therefore, automatic classification systems of plankton images are required to address the huge amounts of images [9].
Currently, some systems have been developed for plankton image classification [10]. Imaging in situ Tang et al. [11] designed a recognition system combining moment invariants and Fourier descriptor with granulometric features using learning vector quantization neural network to classify plankton images detected by VPR; then Hu and Davis [12] improved the classification system with co-occurrence matrices (COM) as the feature and a Support Vector Machine (SVM) as the classifier. Luo et al. [13, 14] presented a system to recognize underwater plankton images from SIPPER, by combining invariant moments and granulometric features with some specific features (such as size, convex ratio, transparency ratio, etc.), and using active learning in conjunction with SVM; and Tang et al. [15] applied shape descriptors and a normalized multilevel dominant eigenvector estimation (NMDEE) method to select a best feature set for binary plankton image classification; then Zhao et al. [16] improved the binary SIPPER plankton image classification using random subspace. Sosik and Olson [17] developed an approach that relies on extraction of image features, including size, shape, symmetry, and texture characteristics, plus orientation invariant moments, diffraction pattern sampling, and co-occurrence matrix statistics, which are then presented to a feature selection and SVM algorithm for classification of images generated by IFCB. Bi et al. [18] also developed a semi-automated approach to analyze plankton taxa from images acquired by ZOOVIS. Faillettaz et al. [19] post-processed the computer-generated classification for images collected by ISIIS using Random Forest (RF) obtained with the ZooProcess and PkID toolchain [8] developed for ZooScan to describe plankton distribution patterns. Imaging in the lab ADIAC [20], which stands for Automatic Diatom Identification And Classification, integrated the shape and texture features with Decision Tree (DT), Neural Network (NN), k Nearest Neighbor (kNN) and ensemble learning methods for diatom recognition [21–23]; Dimitrovski et al. [24] presented a hierarchical multi-label classification (HMC) system for diatom image classification evaluated on the ADIAC [20] database. DiCANN [25] developed a machine learning system for Dinoflagellate Categorisation by Artificial Neural Network. Gorsky et al. [8] presented a semi-automatic approach that entails automated classification of images followed by manual validation within ZooScan integrated system. Bell and Hopcroft [26] assessed ZooImage software with the bundled six classifiers (LDA, RPT, kNN, LVQ, NN, and RF) for the classification of zooplankton. Mosleh et al. [27] developed a freshwater algae classification system by using Artificial Neural Network (ANN) with extracted shape and texture features. Santhi et al. [28] identified algal from microscopic images by applying ANN on extracted and reduced features such as texture, shape, and object boundary. Verikas et al. [29] exploited light and fluorescence microscopic images to extract geometry, shape and texture feature sets which were then selected and used in SVM as well as RF classifiers to distinguish between Prorocentrum minimum cells and other objects.
Analysis of the aforementioned methods shows the performance of plankton image classification systems based on applied features and classifiers, among which the general features, such as size, invariant moments, co-occurrence matrix, Fourier descriptor, etc., and the traditional classifiers, such as SVM, RF, ANN, etc., are most commonly used respectively [8, 11–13, 17, 20, 25, 27, 29]. However, these features usually suffer from robustness shortage and cannot represent the biomorphic characteristics of plankton well. Also the traditional classifiers usually have not high prediction accuracy on different datasets especially more than 20 categories so that they are hard to be directly applied for ecological studies [8, 18, 19]. Recently, with the development of computer vision technologies, some image features (descriptors) have been developed, such as Histograms of Oriented Gradients (HOG), Scale-Invariant Feature Transform (SIFT), Shape Context (SC), Local Binary Pattern (LBP), etc., and they have been proven to be robust against occlusion and clutter, also have a good performance on object detection, recognition and classification [30]. Thus, we think that it’s the time to apply these new robust image descriptors to represent the characteristics of plankton for better classification performance.
In addition, the morphological characteristics of plankton can be described from different views with diverse features, such as shape, gray, texture, etc. [17, 27]. However, directly concatenating all the features into one that is fed to a single learner doesn’t guarantee an optimum performance [31], and it may exacerbate the “curse of dimensionality” [32]. Therefore, we consider that multiple kernel learning (MKL), where different features are fed to different classifiers, might be helpful and necessary to make better use of the information and improve the plankton image classification performance.
Furthermore, the literature of plankton image classification shows that most methods are developed for the specific imaging device and only address a relatively narrow taxonomic scope. Nevertheless, for the abundant species in a wide taxonomic scope from phytoplankton to zooplankton located in all over the world [33], it’s really impossible to design one specific classification system for each application.
In this paper, inspired by the analysis of literature and development of technology, we focus on the requirements of practical application and propose an automatic system for plankton image classification combining multiple view features via multiple kernel learning. On one thing, in order to describe the biomorphic characteristics of plankton more completely and comprehensively, we combine the general features with the latest robust features, especially by adding features like Inner-Distance Shape Context (IDSC) for morphological representation. On the other hand, we divide all the features into different types from multiple views and feed them to multiple classifiers instead of only one by combining different kernel matrices computed from different types of features optimally via multiple kernel learning. Moreover, we also apply feature selection method to choose the optimal feature subsets from redundant features for satisfying different datasets from different imaging devices. We implement our proposed classification system on three different datasets across more than 20 categories from phytoplankton to zooplankton. The experimental results validate that our system outperforms state-of-the-art systems for plankton image classification in terms of accuracy and robustness.
Methods
The automatic plankton image classification we proposed consists of five parts as follows: 1) image pre-processing, 2) feature extraction, 3) feature selection, 4) multiple kernel learning, and 5) evaluation. The framework is shown in Fig. 1.
Image pre-processing
Images captured by (especially in situ) imaging devices mostly suffer from noise (Fig. 2 a). They may contain uninterested regions or unavoidable marine snow. To enhance the image quality and highlight the image features, we implement image pre-processing firstly to extract the plankton cells while reduce the noise such as marine snow in our system.
The image pre-processing operation is the only part that may differ depending on the dataset, because the images acquired by different devices from different samples or locations are usually different in terms of noise and quality. But the objective and result of this operation are the same, that is, to extract the plankton cells with biomorphic characteristics from the original images. In our study, we focused on three different datasets acquired by IFCB, ZooScan, and ISIIS respectively, and designed the following unified steps: 1) binarization: convert the gray scale images to binary images (Fig. 2 b) based on threshold methods, 2) denoising: remove small connected regions (i. e., less than 5 pixels) due to the priori that they might not be plankton cells by morphological operations to obtain the binary mask (Fig. 2 c), and 3) extraction: extract the plankton cells (Fig. 2 d) from the original image using the denoised binary mask.
Feature extraction
To obtain comprehensive characteristics of plankton, we extract various types of features in our classification system, including general features, which have been used for plankton classification previously, and robust features that are used extensively in object detection and recognition currently. The following will introduce our extracted features.
Geometric and grayscale features
Geometric features include size and shape measurements, such as area, circularity, elongation, convex rate, etc., and grayscale features include sum, mean, standard deviation, etc., and these features can describe the basic morphological characteristics of plankton and have been used in the previous study [17, 27, 29]. In our system, the geometric and grayscale features we applied consist of 43 elements represented by a 43-dimensiontal feature vector.
Texture features
Texture is one of the important characteristics used in plankton identification [17, 27]. In our system, we applied four method for texture feature extraction, including Gabor filter, variogram function, Local Binary Pattern (LBP), and Binary Gradient Contours (BGC).
Gabor filter Frequency and orientation representations of Gabor filters, which are similar to those of human visual system, are appropriate for texture representation [34]. In the spatial domain, a 2D Gabor filter is a Gaussian kernel function. The impulse response of these filters is created by convoluting a Gaussian function
where θ represents the orientation, F represents the center frequency, and σ is the standard deviation. Gabor filter is an essentially convolution of original image
where I(x,y) is the original image, Q(x,y) is the Gabor filter result. The mean value and standard deviation of Gabor filter result can be used to describe the texture feature
where m,n represent the size of image. A set of Gabor filters with different frequencies and orientations will be helpful for description of characteristics completely. In our system, we use Gabor filters with 6 kinds of frequencies and 8 kinds of orientations for plankton texture representation as shown in Fig. 3. Therefore, we obtained 48 mean values and standard deviation values to construct a 96-dimentional feature vector.
Variogram function The variogram, which is the basic function in geostatistics, is widely used for extraction of texture characteristics. The mathematical expression of variogram is
where h is certain lag, N(h) is the number of experimental pairs, and I(x),I(x+h) are pixel values at x,x+h. In our system, we applied variogram γ to describe texture features.
Local binary pattern LBP is a classical texture descriptor designed for classification and recognition, especially face recognition [35]. The basic idea of LBP is that two-dimensional surface textures can be described by local spatial patterns and gray scale contrast. The original LBP algorithm labels each pixel of image with 8-bit binary codes called LBP labels, which are obtained by the local structure (i.e., neighborhood) around the pixel. The histogram of these LBP labels can be used as texture descriptor. In our study, we improved the original LBP descriptor by segmenting the image into cells and then concatenating all the cell-based histograms as shown in Fig. 4, which can represent the part-based biomorphic features well.
Binary gradient contours BGC [36] relies on computing the gradient between pairs of pixels all along a closed path around the central pixel of a grayscale image patch. The most frequently used paths are single-loop, double-loop and triple-loop. And the binary gradient contours of single-loop are expressed as
Then all grayscale patterns can be mapped to the binary gradient contour value of single-loop by
The texture is described by the histogram that quantify the occurrence of BGC value in images. In our system, we used the single-loop BGC descriptor.
Granulometric feature
Granulometry [37] is an approach to measure the size distribution of grains in binary image. It describes the particles range and distribution using a series of opening operators with structuring elements of increasing size
where B denotes binary image, Ψ λ (B) denotes the result binary image, T is structuring elements, ∘ means opening operation and λ is the number of opening operation times. The granulometric size distribution of B is given by
where v(B) indicates the pixel number of grains. Therefore, granulometry can represent the multiscale shape feature of object. In our system, we used granulometry to describe the shape feature of plankton with two different setups: one is the size of elements increasing from 2 to 50 at interval of 4, and the other is the size of elements increasing from 5 to 60 at interval of 5.
Local features
Local features refer to patterns or distinct structures found in an image, such as points, edges, etc., and they can describe local image structures while handle scale changes, rotation as well as occlusion.
Histograms of oriented gradients HOG [38] counts occurrences of gradient orientation in localized portions of image. The main idea is that local object appearance and shape within an image can be described by the distribution of intensity gradients or edge directions, e.g., Fig. 5. In our system, every image is processed into square and resized to 256×256, then it is decomposed into 32×32 cells, and our HOG feature descriptor is constructed by the concatenation of the histograms of gradient directions of these cells.
Scale-Invariant feature transform SIFT [39] is a well-known robust algorithm to detect and describe local features of image against scale and rotation changes by extracting keypoints (Fig. 6) and computing their descriptors, which has been widely used for object recognition, robotic mapping, video tracking, and so on. In image classification, SIFT usually integrates with bag-of-words (BoW) model to treat image features as words: first, use SIFT to extract keypoints of all images in dataset; second, divide all keypoints into groups by K-means clustering with codewords as the centers of learned clusters; then, the keypoints in an image can be mapped to a certain codeword through the clustering and an image can be represented by the n-bin histogram of the codewords. In our system, we set the number of clusters to 100, and every image is described by a 100-dimensional feature vector.
Inner-Distance shape context IDSC [40] is extended from Shape Context (SC) [41] designed for shape representation that describes the relative spatial distribution (distance and orientation) of landmark points around feature points. Given n sample points P={p 1,⋯,p n } on a shape, the shape context at point p i is defined as a histogram h i of the relative coordinates of the remaining n−1 points
where the bins uniformly divide the log-polar space. Shape context can be applied to shape matching by calculating the similarity between two shapes. The cost of matching two points p i ,q j is computed by
The matching π should minimize the match cost H(π) defined as
Once the best matching is found, the matching cost H(π) is the similarity measurement between shapes, that is, the shape distance. The shape context uses the Euclidean distance to measure the spatial relation between landmark points, which may cause less discriminability for complex shapes with articulations. The inner-distance, defined as the length of the shortest path within the shape boundary, is a natural way to solve this problem since it captures the shape structure better than Euclidean distance. In our system, we applied IDSC-based shape matching to describe shape features of plankton as follows: first, pick three images from each category of dataset and manually extract their shapes as templates; second, use IDSC to match shape of every image with templates and compute the shape distances between them; then, obtain the shape distances as the feature vector for shape representation.
Feature selection
In machine learning, feature selection is an important process of selecting the optimal features for classification, because the redundant features can suppress the performance of classifier. Besides, feature selection can reduce training time and improve the efficiency, especially for high-dimensional features. Thus, we applied wrapper method [42] for feature selection to choose the optimal features from aforementioned various high-dimensional features for performance improvement. In our system, we divided all features into ten types (Fig. 1), and we applied feature selection on each type of features to choose the optimal features respectively.
Multiple kernel learning
Multiple kernel learning (MKL) is a set of machine learning methods that use a predefined set of kernels and learn an optimal linear or non-linear combination of multiple kernels. It can be applied to select for an optimal kernel and parameters, and combine different types of features. Recently, MKL has received great attention and been used in many recognition and classification applications, such as visual object recognition [43] and hyperspectral image classification [44]. MKL aims to learn a function of the form
with M multiple kernels instead of a single one
where f η is the combination function of kernels, and it can be a linear or non-linear function.
According to the functional form of combination, the existing MKL algorithms can be grouped into three basic categories [31]: 1) linear combination methods, such as SimpleMKL [45], GLMKL [46], 2) nonlinear combination methods, such as GMKL [47], NLMKL [48], and 3) data-dependent combination methods, such as LMKL [49]. Gönen and Alpaydin [31] also performed experiments on real datasets for comparison of existing MKL algorithms and gave an overall comparison between algorithms in terms of misclassification error. It concluded that using multiple kernels is better than using a single one and nonlinear or data-dependent combination seem more promising. Based on their experiments and our analysis, in our system, we chose NLMKL [48], a nonlinear combination of kernels, as MKL method to combine multiple extracted plankton features. NLMKL is based on a polynomial combination of base kernels shown as
We used NLMKL to combine three kernel functions, Gaussian kernel, polynomial kernel, and linear kernel, on each type of features (Fig. 1).
Evaluation
A confusion matrix (Table 1) is a table containing information about actual and predicted classifications, so that it can be used to evaluate the performance of classification systems. Each column of confusion matrix represents the samples in a predicted class while each row represents the samples in an actual class. And the diagonal of the matrix represents correct identifications of samples. Several measures can be derived from a confusion matrix, for instance, true positive rate (TPR, also called recall), false negative rate (FNR), false positive rate (FPR), true negative rate (TNR), positive predictive value (PPV, also called precision). In our system, we use Recall and Precision (actually 1−Precision, means error rate) to evaluate the performance of classification
where True positive is the number of samples correctly predicted, and Condition positive is total number of actual samples. And higher R with lower 1−P will give better classification performance. Then we use F measure (higher) that combines precision and recall with harmonic mean to evaluate the performance (better) of classification system
Results
To illustrate our proposed plankton image classification system, we perform three experiments on three publicly available datasets collected by different imaging devices in different locations with more than 20 categories covering phytoplankton and zooplankton.
Datasets
WHOI dataset
This dataset was collected with IFCB [6] from Woods Hole Harbor water. All sampling was done between late fall and early spring in 2004 and 2005 [17] and can be accessed at: http://onlinelibrary.wiley.com/doi/10.4319/lom.2007.5.204/full. It contains 6600 images with distribution across 22 categories (Fig. 7), and most categories are phytoplankton taxa at the genus level, among which 16 categories are diatoms: Asterionellopsis spp., Chaetoceros spp., Cylindrotheca spp., Cerataulina spp. plus the morphologically similar species of Dactyliosolen such as D. fragilissimus, other species of Dactyliosolen morphologically similar to D. blavyanus, Dinobryon spp., Ditylum spp., Euglena spp. plus other euglenoids, Guinardia spp., Licmophora spp., Phaeocystis spp., Pleurosigma spp., Pseudonitzschia spp., Rhizosolenia spp. and rare cases of Proboscia spp., Skeletonema spp., Thalassiosira spp. and similar centric diatoms; the remaining categories are mixtures of morphologically similar particles and cell types: ciliates, detritus, dinoflagellates greater than 20μ m, nanoflagellates, other cells less than 20μ m, and other single-celled pennate diatoms. The images were split between training and testing sets of equal size, and each set contains 22 categories with 150 individual images in each. Accordingly, in our experiments, we used the training set for learning and the testing set to assess the performance of the classification system.
ZooScan dataset
This dataset was collected by the ZooScan system (http://www.zooscan.com) with a series of samples from the Bay of Villefranche-sur-mer, France between 22 August 2007 and 8 October 2008 [8]. It contains 3771 zooplankton images of 20 categories (Fig. 8), among which 14 categories are zooplankton: Limacina, Pteropoda, Penilia, Oithona, Poecilostomatoida, other species of Copepoda, Decapoda, Appendicularia, Thaliaca, Chaetognatha, Radiolaria, Calycophorae, other species of Medusae, and eggs of zooplankton; the remaining categories are non-zooplankton: bubble, fiber, aggregates, dark aggregates, pseudoplankton, and images with bad focus. The number of images in each category is different as shown in Fig. 9. In our experiments on this dataset, we used 2-fold cross validation to evaluate the performance of the classification system.
Kaggle dataset
This dataset was collected between May-June 2014 in the Straits of Florida using ISIIS [7], and was first published on Kaggle (https://www.kaggle.com/c/datasciencebowl) with data provided by the Hatfield Marine Science Center at Oregon State University. It consists of 121 categories ranging from the smallest single-celled protists to copepods, larval fish, and larger jellies. In our experiments, we chose 38 categories (Fig. 10) with more than 100 individual images in each (Fig. 11) to construct a new dataset, among which 35 categories are plankton and 3 categories are non-plankton including atifacts, atifacts edge and fecal pellet. The constructed dataset contains 28748 images, and we used 5-fold cross validation to evaluate the performance of the classification system.
Experiments
We designed three experiments on above three datasets for evaluation of our classification system: first, we built the baseline system for benchmarking state-of-the-art plankton image classification systems; then, we used SVM with three kernels to compare our extracted features with the baseline; at last, we applied NLMKL on our extracted features to compare our final system with SVM system.
Baseline
To illustrate the performance of our proposed system, we should first build a baseline system to benchmark state-of-the-art plankton image classification systems. The baseline system is built as follows: 1) feature extraction: extract the 210 features used by Sosik and Olson [17] and the 53 features used in ZooScan system [8] to construct a 263-dimensional feature vector, 2) feature selection: apply the feature selection algorithm used in [17] to choose optimal features while remove redundant features and obtain 99, 80, and 100 dimensions of features for WHOI, ZooScan, and Kaggle datasets respectively, and 3) classifier: use SVM with Gaussian kernel to train the classifier and select the optimal C and gamma by searching over the grid of appropriate parameters.
The classification results of baseline system on three datasets are listed in Table 2 and the confusion matrices are shown in Fig. 12. It can be seen that the best classification performance of three datasets are 88.27% Recall with 11.63% 1−Precision, 80.6% Recall with 16.3% 1−Precision, and 75.36% Recall with 21.49% 1−Precision, respectively. The performances on WHOI dataset and ZooScan dataset are a little better than that of Sosik and Olson [17] with 88% overall accuracy and ZooScan system [8] of 78% Recall with 19% 1−Precision. Obviously, the system, which combines their methods, has better performance. Therefore, this baseline can be used as the benchmark for performance evaluation of our proposed system.
Comparison of features
This experiment is designed for performance evaluation of our extracted features, and it is implemented as follows: 1) feature extraction: extract all the features presented in “Feature extraction” section, 2) feature grouping: group all the features into 10 types that are geometric and grayscale features, Gabor features, variogram features, LBP features, BGC features, granulometry1 features, granulometry2 features, SIFT features, HOG features, and IDSC features, 3) feature selection: employ feature selection method presented in “Feature selection” section on each type of features separately and concatenate the output features to obtain 139, 148, and 233 dimensions of features for WHOI, ZooScan, and Kaggle datasets respectively, and 4) classifiers: use SVM with Gaussian, polynomial, and linear kernels to train the classifiers.
The experimental results on three datasets are listed in Table 3 and the confusion matrices are shown in Fig. 13. It can be seen that the best classification performance with highest F measure on WHOI dataset is 89.57% Recall with 10.3% 1−Precision (Gaussian kernel and C=100), which provides 1.3 percentage points better Recall than baseline, and the best classification performance on ZooScan and Kaggle datasets are 85.39% Recall with 13.2% 1−Precision and 82.41% Recall with 16.33% 1−Precision respectively (Gaussian kernel and C=10), which provide 4.79 and 7.05 percentage points better Recall than baseline. Consequently, the features we extracted are able to show more complete description of plankton characteristics and improve the performance of classification.
Comparison of systems
This experiment is designed for performance evaluation of our proposed system, and it is implemented as follows: 1) feature extraction, 2) feature grouping, and 3) feature selection are the same as the “Comparison of features” experiment, and 4) classifiers: use NLMKL to combine various types of features with one kind of kernel and three kinds of kernel to train the classifiers respectively.
The experimental results on three datasets are listed in Table 4 (one kind of kernel) and Table 5 (three kinds of kernel), and the confusion matrices are shown in Fig. 14. It can be seen that the best classification performance with highest F measure of MKL system with one kind of kernel on three datasets are 89.67% Recall with 10.16% 1−Precision (polynomial kernel and C=10), 86.6% Recall with 9.92% 1−Precision (Gaussian kernel and C=100), and 82.6% Recall with 15.62% 1−Precision (polynomial kernel and C=10), which provide 0.1, 1.21, and 0.19 percentage points better Recall than SVM system, while the best classification performance of MKL system with three kinds of kernel on three datasets are 90% Recall with 9.91% 1−Precision, 88.34% Recall with 9.58% 1−Precision, and 83.67% Recall with 14.49% 1−Precision, which provide 0.33, 1.74, and 1.07 percentage points better Recall than the MKL system with one kind of kernel as well as 1.73, 7.74, and 8.31 percentage points better Recall than baseline, respectively. The results validate that MKL is more effective than SVM for classification, and combining more kinds of kernel can provide better performance than only one kind of kernel.
Discussion
An automated plankton classification system for abundance estimation of different plankton categories is presented in our work. The goal of our work points to develop a system that can be widely used for classification of phytoplankton and zooplankton with higher accuracy for ecological studies. With this aim, we proposed an automatic system for plankton classification combining various types of features from different views using MKL, so that it can make better use of information from multiple views and improve the classification performance in terms of accuracy.
The current study of plankton image classification system mainly focuses on one specific imaging device with several categories of phytoplankton or zooplankton, so that it’s hard to have both high accuracy and wide suitability. In order to broaden the application scope and improve the performance of plankton image classification system, we developed the system with threefold contributions: 1) we extracted features from all conceivable views to describe plankton morphological characteristics more completely and comprehensively, 2) we used MKL to combine different views of features for better “understanding” of extracted information, and 3) we combined linear and nonlinear kernels for MKL to obtain better performance than state-of-the-art systems.
In order to evaluate the performance of our proposed system for plankton image classification, three different datasets were collected and constructed while three experiments were designed and implemented in our study. The three datasets, named WHOI, ZooScan and Kaggle, were collected by three different imaging devices, i.e., IFCB, ZooScan, and ISIIS, respectively. The images were sampled in different locations and cover wide categories more than 20 from phytoplankton to zooplankton. The baseline experiment was designed and implemented on the three datasets with state-of-the-art plankton image classification methods to give a benchmark for evaluation. The comparison experiments of features and systems were designed to validate the effectiveness and robustness of features and systems respectively. And the experimental results show that our multiple view features performs better than state-of-the-art and our MKL system combing these multiple view features performs best in all the experiments.
Tables 2 and 3 illustrate the accuracy of the first and second experiments. In comparison, it can be seen that our proposed multiple view features are helpful to improve the plankton classification accuracy, since that the best classification performance of Recall on three datasets increase by 1.3% (WHOI), 4.79% (ZooScan), and 7.05% (Kaggle), and meanwhile the corresponding error rates drop by 1.33% (WHOI), 3.1% (ZooScan), and 5.16% (Kaggle), respectively. And by comparing the confusion matrices shown in Figs. 12 and 13, we can find that the classification accuracy of 14 categories in WHOI dataset, 16 categories in ZooScan dataset, and 36 categories in Kaggle dataset are improved in the second experiment. These results show that our proposed multiple view features can describe the plankton morphological characteristics more completely and comprehensively, for example, as shown in Fig. 12, 58 images of “Appendicularia” category in ZooScan dataset are incorrectly classified as “Gelatinous_sipho_cloche” in the baseline experiment, but this misclassified number drops from 58 to 10 in the second experiment using our multiple view features shown in Fig. 13.
With the same multiple view features, the second and third experiments use different machine learning strategies for classification. By comparing Table 3 with Table 4, we can observe that MKL provides better performance than SVM. NLMKL with one kind of kernel in the third experiment provides 0.1 (WHOI), 1.21 (ZooScan), and 0.19 (Kaggle) percentage points better Recall than SVM, and meanwhile the corresponding error rates drop by 0.14% (WHOI), 3.28% (ZooScan), and 0.71% (Kaggle), respectively. And by comparing the confusion matrices shown in Figs. 13 and 14, there are 11 categories in WHOI dataset, all 20 categories in ZooScan dataset, and 21 categories in Kaggle dataset having higher accuracy in the third experiment. By comparing Table 4 with Table 5, we can find that MKL with more kernels performs better than only one. In other words, MKL, which optimally combines different kernel matrices computed from multiple types of features with multiple kernel functions, can make better use of information of each type of features and improve the classification performance significantly. Not surprisingly, MKL has been proven to be a useful tool for feature combination to enhance the discrimination power of classifiers, for example, Althloothi et al. [50] improved the performance of human activity recognition by combining two sets of features using MKL, and Luo et al. [51] proved the effectiveness of MKL for feature combination.
Our overall results compare favorably with previous results. Our proposed plankton image classification system has a classification performance of 90% Recall on WHOI dataset, 88.34% Recall on ZooScan dataset, and 83.67% Recall on Kaggle dataset, respectively, and meanwhile the corresponding error rates have the similar degree of decline. It is effective for situations like these three datasets. By comparing the confusion matrices shown in Figs. 12 and 14, it can be seen that most categories of the three datasets have higher accuracy using our system than baseline with state-of-the-art systems. Based on these comparisons, we can conclude that multiple view features combination using MKL enhances the performance of classification and accommodates a wide variety of plankton categories while gives significantly better accuracy.
There is one important issue that we should discuss in this paper: deep learning for plankton image classification. Deep learning [52], also known as deep neural network, has been drawn more and more attention these years because of its dramatically performance on large-scale visual recognition and classification tasks. It attempts to model high level abstractions in data in order to learn representations of data end-to-end. This technology has also been used in plankton image classification recently [53–55]. However, currently, deep learning can perform good only when fed with enough human labeled data. That’s why it is strongly related to the word “large-scale”. With the development of in situ underwater microscopy imaging technology, it’s easy to collect thousands and thousands of plankton images in a short time (such as one day), but it’s still very hard to label all of them by experts because it needs very professional knowledge to identify them. Therefore, it would be very useful to build an end-to-end learning system with very less labeled data, which is also an important future work of deep learning field. The system proposed in this paper doesn’t need large-scale data and performs good on classification for more than 20 categories (but less than 50 categories), which is more suitable for the current actual applications of ecological studies.
Conclusions
In this paper, we propose an automatic plankton image classification system combining multiple view features using multiple kernel learning. In our system, multiple view features, including general features and robust features, are extracted for better describing the morphological characteristics of plankton. And different types of features are combined via NLMKL using three kernel functions (linear, polynomial and Gaussian kernel functions) in order to use information of features better and achieve a higher classification rate. Our experimental results on three different datasets show the significance of the extracted multiple view features and MKL in plankton image classification. The main limitation of our system is lack of ability to work well with terribly imbalanced datasets, and we wish to leave this part in our future work.
To encourage future work, we make all the source code and datasets available in the GitHub repository: https://github.com/zhenglab/PlanktonMKL.
Abbreviations
- ADIAC:
-
Automatic diatom identification and classification
- ANN:
-
Artificial neural network
- BGC:
-
Binary gradient contours
- BoW:
-
Bag-of-words
- COM:
-
Co-occurrence matrices
- DiCANN:
-
Dinoflagellate categorisation by artificial neural network
- DT:
-
Decision tree
- FNR:
-
False negative rate
- FPR:
-
False positive rate
- GLMKL:
-
Group Lasso-based MKL
- GMKL:
-
Generalized MKL
- HMC:
-
Hierarchical multi-label classification
- HOG:
-
Histograms of oriented gradients
- IDSC:
-
Inner-distance shape context
- IFCB:
-
Imaging FlowCytoBot
- ISIIS:
-
In Situ Ichthyoplankton imaging system
- kNN:
-
k Nearest neighbor
- LBP:
-
Local binary pattern
- LDA:
-
Linear discriminant analysis
- LMKL:
-
Localized MKL
- LVQ:
-
Learning vector quantization
- MKL:
-
Multiple kernel learning
- NLMKL:
-
Non-Linear MKL
- NMDEE:
-
Normalized multilevel dominant Eigenvector estimation
- NN:
-
Neural network
- PPV:
-
Positive predictive value
- RF:
-
Random forest
- RPT:
-
Recursive partitioning tree
- SC:
-
Shape context
- SIPPER:
-
Shadowed image particle profiling evaluation recorder
- SIFT:
-
Scale-invariant feature transform
- SPC:
-
Scripps plankton camera
- SVM:
-
Support vector machine
- TNR:
-
True negative rate
- TPR:
-
True positive rate
- UVP:
-
Underwater video profiler
- VPR:
-
Video plankton recorder
- ZOOVIS:
-
ZOOplankton VIsualization system
References
Davis CS, Thwaites FT, Gallager SM, Hu Q. A three-axis fast-tow digital Video Plankton Recorder for rapid surveys of plankton taxa and hydrography. Limnol Oceanogr Meth. 2005; 3:59–74.
Picheral M, Guidi L, Stemmann L, Karl DM, Iddaoud G, Gorsky G. The Underwater Vision Profiler 5: An advanced instrument for high spatial resolution studies of particle size spectra and zooplankton. Limnol Oceanogr Meth. 2010; 8:462–73.
Samson S, Hopkins T, Remsen A, Langebrake L, Sutton T, Patten J. A system for high-resolution zooplankton imaging. IEEE J Oceanic Eng. 2001; 26:671–6.
Benfield MC, Schwehm CJ, Keenan SF. ZOOVIS: a high resolution digital camera system for quantifying zooplankton abundance and environmental data. Proc Am Soc Limnol Oceanogr. 2001;12–17.
Jaffe JS. To sea and to see: That is the answer. Meth Oceanogr. 2016; 15:3–20.
Olson RJ, Sosik HM. A submersible imaging-in-flow instrument to analyze nano-and microplankton: Imaging FlowCytobot. Limnol Oceanogr Meth. 2007; 5:195–203.
Cowen RK, Guigand CM. In situ ichthyoplankton imaging system (isiis): system design and preliminary results. Limnol Oceanogr Meth. 2008; 6:126–32.
Gorsky G, Ohman MD, Picheral M, Gasparini S, Stemmann L, Romagnan JB, Cawood A, Pesant S, García-Comas C, Prejger F. Digital zooplankton image analysis using the ZooScan integrated system. J Plankton Res. 2010; 32:285–303.
MacLeod N, Benfield M, Culverhouse P. Time to automate identification. Nat. 2010; 467:154–5.
Benfield MC, Grosjean P, Culverhouse PF, Irigoien X, Sieracki ME, Lopez-Urrutia A, Dam HG, Hu Q, Davis CS, Hansen A, Pilskaln CH, Riseman EM, Schultz H, Utgoff PE, Gorsky G. RAPID: research on automated plankton identification. Oceanogr. 2007; 20:172–87.
Tang X, Stewart WK, Vincent L, Huang H, Marra M, Gallager SM, Davis CS. Automatic plankton image recognition. Artif Intell Rev. 1998; 12:177–99.
Hu Q, Davis C. Automatic plankton image recognition with co-occurrence matrices and support vector machine. Mar Ecol Prog Ser. 2005; 295:21–31.
Luo T, Kramer K, Goldgof DB, Hall LO, Samson S, Remsen A, Hopkins T. Recognizing plankton images from the shadow image particle profiling evaluation recorder. IEEE Trans Syst Man Cybern B. 2004; 34:1753–62.
Luo T, Kramer K, Goldgof DB, Hall LO, Samson S, Remsen A, Hopkins T. Active learning to recognize multiple types of plankton. J Mach Learn Res. 2005; 6:589–613.
Tang X, Lin F, Samson S, Remsen A. Binary plankton image classification. IEEE J Oceanic Eng. 2006; 31:728–35.
Zhao F, Lin F, Seah HS. Binary SIPPER plankton image classification using random subspace. Neurocomputing. 2010; 73:1853–60.
Sosik HM, Olson RJ. Automated taxonomic classification of phytoplankton sampled with imaging-in-flow cytometry. Limnol Oceanogr Meth. 2007; 5:204–16.
Bi H, Guo Z, Benfield MC, Fan C, Ford M, Shahrestani S, Sieracki JM. A semi-automated image analysis procedure for in situ plankton imaging systems. PLoS ONE. 2015; 10:0127121.
Faillettaz R, Picheral M, Luo JY, Guigand C, Cowen RK, Irisson JO. Imperfect automatic image classification successfully describes plankton distribution patterns. Meth Oceanogr. 2016; 15:60–77.
Du Buf H, Bayer MM. Automatic Diatom Identification. Singapore: World Scientific; 2002.
Loke RE, du Buf JH, Bayer M, Mann DG. Diatom classification in ecological applications. Pattern Recogn. 2004; 37:1283–5.
Jalba AC, Wilkinson MH, Roerdink JB, Bayer MM, Juggins S. Automatic diatom identification using contour analysis by morphological curvature scale spaces. Mach Vis Appl. 2005; 16:217–28.
Hicks YA, Marshall D, Rosin PL, Martin RR, Mann DG, Droop SJM. A model of diatom shape and texture for analysis, synthesis and identification. Mach Vis Appl. 2006; 17:297–307.
Dimitrovski I, Kocev D, Loskovska S, Džeroski S. Hierarchical classification of diatom images using ensembles of predictive clustering trees. Ecol Inform. 2012; 7:19–29.
Culverhouse PF, Herry V, Ellis R, Williams R, Reguera B, Gonzalez-Gil S, Umani SF, Cabrini M, Parisini T. Dinoflagellate categorisation by artificial neural network. Sea Technol. 2002; 43:39–46.
Bell JL, Hopcroft RR. Assessment of ZooImage as a tool for the classification of zooplankton. J Plankton Res. 2008; 30:1351–67.
Mosleh MA, Manssor H, Malek S, Milow P, Salleh A. A preliminary study on automated freshwater algae recognition and classification system. BMC Bioinformatics. 2012; (Suppl 17):25.
Santhi N, Pradeepa C, Subashini P, Kalaiselvi S. Automatic identification of algal community from microscopic images. Bioinform Biol Insights. 2013; 7:327–34.
Verikas A, Gelzinis A, Bacauskiene M, Olenina I, Vaiciukynas E. An integrated approach to analysis of phytoplankton images. IEEE J Oceanic Eng. 2015; 40:315–26.
Fan B, Wang Z, Wu F. Local Image Descriptor: Modern Approaches. Berlin: Springer; 2015.
Gönen M, Alpaydın E. Multiple kernel learning algorithms. J Mach Learn Res. 2011; 12:2211–268.
Bellman RE. Adaptive Control Processes: a Guided Tour. Princeton: Princeton University Press; 2015.
Jouenne F, Probert I, Vaulot D. Plankton taxonomy in the computer age. Cah Biol Mar. 2008; 49:355–67.
Idrissa M, Acheroy M. Texture classification using Gabor filters. Pattern Recogn Lett. 2002; 23:1095–102.
Ahonen T, Hadid A, Pietikainen M. Face description with local binary patterns: Application to face recognition. IEEE Trans Pattern Anal Mach Intell. 2006; 28:2037–41.
Fernández A, Álvarez MX, Bianconi F. Image classification with binary gradient contours. Opt Laser Eng. 2011; 49:1177–84.
Matheron G. Randoms Sets and Integral Equation. New York: Wiley; 1978.
Dalal N, Triggs B. Histograms of oriented gradients for human detection. In: Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition. San Diego: IEEE;2005. p. 886–93.
Lowe DG. Object recognition from local scale-invariant features. In: Proceedings of IEEE International Conference on Computer Vision. Kerkyra: IEEE;1999. p. 1150–57.
Ling H, Jacobs DW. Shape classification using the inner-distance. IEEE Trans Pattern Anal Mach Intell. 2007; 29:286–99.
Belongie S, Malik J, Puzicha J. Shape matching and object recognition using shape contexts. IEEE Trans Pattern Anal Mach Intell. 2002; 24:509–22.
Kohavi R, John GH. Wrappers for feature subset selection. Artif Intell. 1997; 97:273–324.
Bucak SS, Jin R, Jain AK. Multiple kernel learning for visual object recognition: A review. IEEE Trans Pattern Anal Mach Intell. 2014; 36:1354–69.
Li J, Huang X, Gamba P, Bioucas-Dias JM, Zhang L, Benediktsson JA, Plaza A. Multiple feature learning for hyperspectral image classification. IEEE Trans Geosci Remote Sens. 2015; 53:1592–06.
Rakotomamonjy A, Bach FR, Canu S, Grandvalet Y. SimpleMKL. J Mach Learn Res. 2008; 9:2491–521.
Xu Z, Jin R, Yang H, King I, Lyu MR. Simple and efficient multiple kernel learning by group lasso. In: Proceedings of International Conference on Machine Learning. Haifa: Omnipress;2010. p. 1175–82.
Varma M, Babu BR. More generality in efficient multiple kernel learning. In: Proceedings of International Conference on Machine Learning. Montreal: ACM;2009. p. 1065–72.
Cortes C, Mohri M, Rostamizadeh A. Learning non-linear combinations of kernels. In: Proceedings of Advances in Neural Information Processing Systems. Vancouver: Curran Associates;2009. p. 396–404.
Gönen M, Alpaydin E. Localized multiple kernel learning. In: Proceedings of International Conference on Machine Learning. Helsinki: ACM;2008. p. 352–9.
Althloothi S, Mahoor MH, Zhang X, Voyles RM. Human activity recognition using multi-features and multiple kernel learning. Pattern Recogn. 2014; 47:1800–12.
Luo W, Yang J, Xu W, Li J, Zhang J. Higher-level feature combination via multiple kernel learning for image classification. Neurocomputing. 2015; 167:209–17.
LeCun Y, Bengio Y, Hinton G. Deep learning. Nat. 2015; 521:436–44.
Dai J, Wang R, Zheng H, Ji G, Qiao X. ZooplanktoNet: Deep convolutional network for zooplankton classification. In: Proceedings of OCEANS MTS/IEEE Shanghai. Shanghai: IEEE;2016. p. 1–6.
Li X, Cui Z. Deep residual networks for plankton classification. In: Proceedings of OCEANS MTS/IEEE Monterey. Monterey: IEEE;2016. p. 1–4.
Lee H, Park M, Kim J. Plankton classification on imbalanced large scale database via convolutional neural networks with transfer learning. In: Proceedings of IEEE International Conference on Image Processing. Phoenix: IEEE;2016. p. 3713–717.
Acknowledgements
Not applicable.
Funding
This work was supported by the National Natural Science Foundation of China [61771440, 41776113, 61301240], and Qingdao Municipal Science and Technology Program [17-1-1-5-jch].
Publication charges for this article have been funded by the Ocean University of China.
Availability of data and materials
The code and datasets used in the paper are available at https://github.com/zhenglab/PlanktonMKL.
About this supplement
This article has been published as part of BMC Bioinformatics Volume 18 Supplement 16, 2017: 16th International Conference on Bioinformatics (InCoB 2017): Bioinformatics. The full contents of the supplement are available online at https://bmcbioinformatics.biomedcentral.com/articles/supplements/volume-18-supplement-16.
Authors’ contributions
BZ headed the study and structured the whole research. HZ and RW performed the experiments and were major contributors in writing the manuscript. ZY, NW and ZG assisted in manuscript writing. All authors read and approved the final manuscript.
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Zheng, H., Wang, R., Yu, Z. et al. Automatic plankton image classification combining multiple view features via multiple kernel learning. BMC Bioinformatics 18 (Suppl 16), 570 (2017). https://doi.org/10.1186/s12859-017-1954-8
Published:
DOI: https://doi.org/10.1186/s12859-017-1954-8