- Research article
- Open Access
Interactive visual exploration of overlapping similar structures for three-dimensional microscope images
- Megumi Nakao†1Email author,
- Shintaro Takemoto†1,
- Tadao Sugiura†2,
- Kazuaki Sawada†3,
- Ryosuke Kawakami†3, 4,
- Tomomi Nemoto†3, 4 and
- Tetsuya Matsuda†1
© Nakao et al.; licensee BioMed Central. 2014
Received: 30 July 2014
Accepted: 9 December 2014
Published: 19 December 2014
Recent advances in microscopy enable the acquisition of large numbers of tomographic images from living tissues. Three-dimensional microscope images are often displayed with volume rendering by adjusting the transfer functions. However, because the emissions from fluorescent materials and the optical properties based on point spread functions affect the imaging results, the intensity value can differ locally, even in the same structure. Further, images obtained from brain tissues contain a variety of neural structures such as dendrites and axons with complex crossings and overlapping linear structures. In these cases, the transfer functions previously used fail to optimize image generation, making it difficult to explore the connectivity of these tissues.
This paper proposes an interactive visual exploration method by which the transfer functions are modified locally and interactively based on multidimensional features in the images. A direct editing interface is also provided to specify both the target region and structures with characteristic features, where all manual operations can be performed on the rendered image. This method is demonstrated using two-photon microscope images acquired from living mice, and is shown to be an effective method for interactive visual exploration of overlapping similar structures.
An interactive visualization method was introduced for local improvement of visualization by volume rendering in two-photon microscope images containing regions in which linear nerve structures crisscross in a complex manner. The proposed method is characterized by the localized multidimensional transfer function and interface where the parameters can be determined by the user to suit their particular visualization requirements.
Given the complex three-dimensional (3D) architecture of the brain, it is essential to explore the morphology and activity of neurons in all layers of the cortex. However, this can often be challenging because the dendrites of cortical neurons are widely spread across several layers, including deeper layers that are difficult to observe by confocal or light microscopy. The length of the dendrites can vary from 20 μm to 1 mm, and the width and branching of the dendrites depends on the distance from the soma. This suggests that spatial differences in brain morphology relate to the functionality of the neuron, including characteristics of the dendrite and synaptic efficiency ,.
To understand the 3D structure of dendrites and the connectivity among neurons in the brain, in vivo imaging and visualization have significant roles. Recently, the improved performance of microscopy systems has enabled the acquisition of large amounts of slice images from living tissues. In comparison with confocal or other optical microscopy systems, two-photon microscopy has an advantage in visualizing the morphology of neurons within deeper layers of a living mouse brain -. Because the structures of tissues are stored as volume data, volume visualization techniques - are focused on the interactive exploration of the 3D images. When visualizing unknown features in the deeper layers of the brain, prior knowledge of the morphology of tissues , cannot be used. Furthermore, microscopic images are affected by optical characteristics such as scattering within tissues and the presence of image noise within the deeper regions of the images. Because large amounts of volume data are obtained through two-photon microscopy, there is a demand for efficient visualization of local internal structures and characteristic intensity distributions.
Direct volume rendering (DVR) has been widely used for visualizing volume data, where the rendered image is generated from the data by simulating optical properties such as radiation and absorption ,. The user can interactively explore the micro-level structures included in the volume data while changing the camera parameters, modifying the transfer functions - or generating the cross-section of the 3D images. Unlike the pattern recognition approach -, visualization does not involve algorithm-based detection for specific objects. In other words, the only aspects defined by the method are the transformations when visualizing the 3D image as a projection on the screen, and modifications of the visualization parameters and final judgments about the structures observed are left to the user.
The final quality of projections obtained with volume rendering depends significantly on the definition of the transfer functions (TFs). For this reason, the design of the TFs is regarded as an important area of research for volume visualization. Clinical tools commonly provide predefined TFs (TF presets) for efficiency in the clinical workflow. Recent studies have improved the TFs based on multidimensional feature values to visualize changes in texture and morphological characteristics included in the images -. Because the high degree of freedom in multidimensional TFs makes it difficult for users to obtain visualization results through manual parameter settings, a variety of user interfaces and automatic TF generation methods have been investigated -. In the case of visual exploration, there are many situations for which feature descriptors have not been formulated . Some researchers have focused on this issue and have investigated methods for exploring high-dimensional feature space. Principal component analysis, independent component analysis, and clustering techniques are commonly used for dimensionality reduction of the feature space ,.
This paper proposes an interactive visual exploration method by which the TFs are modified locally and interactively based on multidimensional features of the images. This method will also provide a direct editing interface to specify both the target region and the structures with characteristic features, where all manual operations can be performed on the rendered image. Therefore, multidimensional features and interactive methods can be used for local improvement of the visualization results for overlapping structures. Despite the high dimensionality of the TFs, users only specify the structures of interest on the rendered image and control simple parameters to explore similar structures. The performance of the method is demonstrated using two-photon microscope images measured from live mice. The experiments show that this method is effective for interactive visual exploration of overlapping similar structures.
Regarding related study in vessel visualization, the recent work by Kubisch et al.  summarizes problems and reviews of related work. Much recent work has focused on multi-scale methods based on eigenvalue analysis of the Hessian . Using the vesselness measure, Lathen et al.  presented automatic tuning techniques by applying a locally shifted intensity to the TFs (called a TF shift) in the vessel visualization domain. In computed tomography angiography (CTA), which is analogous to microscope images, the intensity value of blood vessels partially decrease owing to the distribution deflection of the contrast medium. The advantage of this approach is to use the 1D TF presets that are popular in clinical applications and are easily set up in a visualization workflow.
In this study, we consider an application of the TF shift technique to 3D microscope images. In biological visualization using two-photon or confocal microscope images, however, the main focus is on neural structures that crisscross in a complex way with multiple overlaps. The deep layers of the brain tissue contain a variety of neural structures with complex shapes such as soma, dendrites, and white matter. Unlike the situation that exists in clinical CT or MRI images, however, numerous minute or thin structures are closely observed with optical scattering noise, which creates challenges for volume visualization. To our knowledge there have been no reports on an interactive, visual exploration software and interface for overlapping similar neural structures in microscope images. This study proposes a new TF shift mechanism and interface that can efficiently use more general multidimensional feature values, while avoiding a complex TF design process.
Figure 1 shows slice images that include structures with characteristic shapes measured from a live mouse using a two-photon microscope. The central image shows the results of visualization of the entire image data with an intensity-based 1D TF. Apical dendrites, dendrites, soma, and white matter, which are all parts of the neurons, are included in the images. In the apical dendritic region, linear structures run along the xy plane and each linear structure crisscrosses in a complex manner. The large linear shadow shown within the red frame in the middle of the image indicates blood vessels, because there is a tendency for a decrease in the intensity value of structures directly below blood vessels. In the dendritic regions, linear structures with a high intensity value can be found along the z axis and, around these, linear structures with low intensity values exist in large quantities, though these structures are not visible. The soma can be identified as a spherical structure. In the white matter region, most of the linear structures run in the specific direction of the xy plane.
We will now discuss our goals toward visualization in the microscope images. Figure 1(b) is a tomographic image of apical dendrites located near the brain surface. However, there is actually a large quantity of low luminance neurons around the dendrites, making observation of the dendrites difficult. Blood vessels with a diameter of ~30 voxels run within this data, and are seen as regions in which nerves do not exist. The low contrast area shadowed by the blood vessels is shown within the red frame in Figure 1(b). Figure 1(c) and (d) shows the volume visualization results obtained by a traditional intensity-based 1D TF. The histogram of the voxel and the opacity curve setting in the TF is inserted in the bottom of each figure. In Figure 1(c), it is not possible to distinguish most of the structures with low intensity under the area of blood vessels. Figure 1(d) shows another visualization result obtained by adjusting the window level parameter of the TF to a lower level. In this figure, some structures can be observed in the low contrast area. However, because of a widening in the range of opaque intensity, surrounding dendric structures and optical noises can occlude the target structure, which fails to distinguish the connectivity of the three-dimensional neural networks. Figure 1(e) is a visualization result obtained by the method proposed in this study, which succeeds in distinguishing dendric structures and their connectivity under the blood vessels. Because the color/opacity outside the target area does not change, the target structures with low intensity as well as the other structures outside of the region of interest can be seen simultaneously without self-occlusion or noise enhancement.
Here, we describe a set of methods to achieve the above-mentioned local refinement of the rendered image. Because intensity values and contrast can differ locally in the target data of this study, the method developed by Lathen et al. for CTA images  is used as a basis for the local adjustment of TFs. To achieve automatic tuning of visualization results for linear structures, Lathen et al. employed the vesselness measure as feature values. In our study, for visualization of various neural structures such as soma, dendrites, and apical dendrites with different features and sizes, we design a novel TF shift framework to address these multidimensional feature values. In addition, for efficient visualization of dendric structures that crisscross in a complex manner, locally similar structures are selectively visualized by direct editing of the visualization results on the rendered image.
Localized multidimensional transfer function
where I is the intensity distribution of the volume data.
For selective visualization of a variety of structures contained in a microscope image, the design of the calculation formula of the shift value Δ of the TF becomes important. Specifically, unlike the previous TF shift technique based on the vesselness measure, multidimensional features are addressed in our framework. The following section explains the concrete calculation formula of Δ. In addition, to allow users to interactively input x 0 and x Ω into the system, it is essential to develop a method for inputting 3D coordinates on the rendering image.
If I(x ) = I(x 0), then RGBA(x ) = RGBA(x 0).
If C (x ) ≈ C (x 0), RGBA(x ) ≈ RGBA(x 0).
Only RGBA(x ) changes within the region Ω near x Ω.
In these conditions, the color/opacity value of point x is expressed as RGBA(x ), and the feature value of point x is expressed as C (x ) ∈R n .
Direct editing interface
In this section, we consider the method for directly inputting the 3D coordinates x 0 and x Ω onto the rendered image. With the volume rendering process, volume data are projected in the viewing direction ν eye ∈R 3. As such, there are voxels drawn by the click point (x click ∈R 2) of the rendered image in the direction of ν eye , which originates from x click . Among the quantity of voxels existing in the direction of ν eye , those close to the surface of the structure are believed to be the voxels specified by the user and are estimated using degrees of transparency ,. First, the volume data are searched from the x click point in the direction of ν eye . During the search, voxels are sampled at fixed intervals, x i . Next, the sum of the opacity value (α i ) of x i is calculated. The value of α i is calculated using TF(x , I(x ), Δ(x )), which considers Δ. The voxel is acquired that is larger than a threshold opacity value (α th ) predefined by , and these voxels are assumed to be those specified by the user and are close to the surface of the structure that has been visualized to consider the Δ.
Results and discussion
We implement a sequence of algorithms using C++, OpenGL, GLSL (Open GL Shader Language), and the software package NVIDIA CUDA (Compute Unified Device Architecture). We apply the developed software to two-photon microscope images and verify the visualization of the characteristic structures and the intensity distribution included in the images with biological researchers. For the 3D microscope image and for visualization of its feature volume, we use the texture-based rendering scheme  to achieve high-speed volume rendering using the texture interpolation and synthesis functions of the graphics processing unit. For verification, we used three volume data sets taken from live, genetically modified mice  using a Nikon two-photon microscope (A1MP+), wherein the neurons in the second and fifth layers of the mouse cortex are labeled by a green fluorescent protein. The study was carried out in accordance with the recommendations in the Guidelines for the Care and Use of Laboratory Animals of the Animal Research Committee. The protocol was approved by the Committee on the Ethics of Animal Experiments. These data sets capture a tomogram with a depth ~1.4 mm from the surface layer of the cortex. The volume data have a size of 512 × 512 × 325 voxels, and the range of capture is 512 × 512 × 1300 μm. Neurons are included within the image, and the direction from the deep part of the brain toward the brain surface is assigned as the positive z axis. Apical dendrites, dendrites, soma, and white matter exist in the positive z direction. To reduce noise, a median filter is applied for preprocessing the volume data.
As with e′ 3 and λ ratio , and by using the condition λ 1 ≤ λ 2 ≤ 0, e′ 3 is prevented from having a value in voxels with an intensity less than that of the background.
The TF preset used in Figure 5(a) is used for our method, and the structure already visualized is indicated as a feature point. The x 0 and Ω are set as shown in (c). The results of the interactive visual exploration with α = 10 and β = 6 are shown in (d). Compared with the visualization results obtained with 1D TF presets (Figure 5(a) and (b)), the results obtained with our method (Figure 5(d)) present the existence of structures within the red frame that are easily distinguished by users
Next, we examine the effect of the sensitivity α and the radius β on the visualization results. The parameters are varied so that α = 1, 10, and 100 and β = 1, 10, and 20, and the results of visualization by combining each of these values are shown in Figure 5(e). When α = 1, structures are completely visualized except for the background; when α = 10, linear structures in the (1, −1, 0) direction are visualized; and when α = 100, the visualization results are the same before and after the operation. When the focus is on visualizing similar structures of x 0, a suitable parameter value for α which enables visualization of linear structures in the direction (1, −1, 0) is α = 10. When α is larger than this value, it is not possible to visualize targeted structures, and when α is smaller than this value, structures outside of those targeted are visualized. Next, when β = 1, the visualization results before and after the operation are practically unchanged. When β = 10, both the target structures and surrounding structures are visualized, and when β = 20, structures corresponding to a wider range than β = 10 are visualized. Thus, to visualize the target structures only, it is best to use the smallest β possible. However, if β is too small, the function cannot be used as a guide when specifying the target structures. Conversely, although it is easy to specify target structures if β is large, it is possible that structures outside of those targeted may also be visualized.
As such, it is necessary to select suitable parameters based on user objectives and on the actual situation when attempting to obtain good visualization results. Although the parameters are modulated based on a trial-and-error process, the following guidelines can be considered. Because α controls sensitivity of the features that are to be visualized, large α values are first suitable to roughly visualize local structures with a variety of features, and then smaller values can be tried to strictly extract the target features. β controls the size of the direct editing pointer that defines the minimal region of interest (ROI). Therefore, β can be adjusted based on the size of the target structures. Larger β is useful when searching non-visualized, transparent structures existing within the volume data. After visualizing a part of the target structure, a small β is useful for fine adjustment of the visualization results.
where the quantity σ is the standard deviation of λ ratio .
The point of the linear structures is denoted as x 0. To apply small modifications to the visualization results, we use (α, β) = (0.30, 6). By changing the specified region of Ω, we interactively obtain the visualization result. First, clear structures are visualized among the linear structures when the dendrites within the area B are taken as the starting point. The proposed method enables the presentation of plural structures in forms distinguishable to the eye, which are difficult to observe in semitransparency with methods developed to date. In addition, when using the vesselness measure based TF shift, many overlapping structures are displayed simultaneously. However, the proposed method allows the structures of interest to be traced easily on the rendered image. Figure 7(d) shows the visualization results after interactive editing of the dendric structures. The TF shift is shown in Figure 7(e). Comparing the two methods, linear structures that are largely transparent and cannot be distinguished can be visualized using the proposed method.
Figure 8(b) shows the MIP-based DVR results in the Vaa3D software. This rendering mode is effective for visualizing occluded vessels, and some neural structures can actually be observed in the lower intensity area. However, because of a widening in the range of opaque intensity, optical noise or unfocused surrounding structures are also visualized globally, which can decrease the contrast between the neural structures and background areas. This continues to make it difficult to clearly visualize connectivity of the neural structures. Figure 8(c) shows the Vaa3D-Neuron2 auto-tracing result with an auto-thresholding mode . Neural structures are reconstructed as surface models by correcting imperfect parts in the traced areas. However, this function can yield many artifacts, especially around complex crossing structures with low-contrast intensity values. When higher threshold values are used, the artifacts can be reduced but reconstruction of the neural structures in low intensity area is not achieved. This results in a difference in the spatial density of neural structures in low-intensity areas and in other areas. We also note that the tracing result is represented using the reconstructed surface models. Locally improved DVR is not achieved using only vessel tracing tools.
With the proposed method, the TF shift is applied to the low-intensity area locally and interactively. The parameters used in this experiment are (α, β) = (0.45, 20). Figure 8(d) shows the spatial relationship between the TF shift applied to a low-intensity area and other dendric structures with high intensity. Figure 8(e) is the final rendering result, and the connectivity of the dendric structures can be viewed thanks to locally improved DVR images. Compared with (a), (b) and (c) of Figure 8, the visual appearance of the spatial density for vessel structures is corrected, which generates a more natural visualization result in (e). Thus, our approach can simultaneously visualize regions that are below blood vessels without affecting regions that are not.
So far, most interaction techniques ,, assume that the entire shapes of the target structures is first visualized in the screen. Then, by pinpointing a 3D position on the rendered image, the user can extract a part of the 3D structure to specify his/her ROI. However, microscopic images obtained from brain tissues sometimes contain a variety of neural structures with complex crossings and overlapping linear structures. In these cases, even if the 3D location is specified, the tracing task is difficult and often fails to extract a single linear structure because undesired surrounding structures are simultaneously selected. In the developed software, we can start the tracing task on the partially-visualized structure and can expand it to transparent regions with low-intensity areas where no structures are yet visualized (see Figure 3). In addition, thanks to the multidimensional TF, the user can limit the visible structures in a robust way based upon a combination of 3D features such as direction, radius and textures (Figures 6, 7 and 8). The multidimensional TF design and its combination with the direct editing interface are main contributions of this work in biological visualization. In addition, three-dimensional images with scalar intensity values measured from electron microscopes and confocal microscopes as well as CT/MRI images can be applied to our software. We will further study practical application to other types of volume data
The automation of transfer function (TF) generation is a critical issue in volume rendering. Specifically, in biological volume visualization, users often focus on overlapped neural structures or internal structures occluded by other surrounding tissues. When automatic TF tuning , is applied globally, many structures are simultaneously visualized and some of these structures occlude the target structures. In Figure 7(c), because similar structures in the data are highlighted, it is difficult to present the endpoints of one linear structure in a visually distinguishable form. Applying machine learning or clustering approaches  are interesting methods to automate TF configuration, but they require prior knowledge to learn the target structures.
Knowledge exploration for newly-measured data is regarded as a user-dependent problem in many situations. The advantage of the interactive, manual TF configuration is that features of interest can be locally explored based on the user's preferences. This user-dependent localization of information visualization is an important factor to transfer the user's biological knowledge into the visualization system. Also, because the developed system does not require any time-consuming setup, it can provide a practical environment for rapid visualization of the measured data. Combining the interactive localization with automatic TF tuning would be interesting and has the possibility of a more intelligent interface for volume visualization, and we would like to further study this on the semi-automatic framework.
In this study, a method is proposed for interactive and local improvement of visualization by volume rendering in two-photon microscope images containing regions in which linear nerve structures crisscross in a complex manner. For this method, multidimensional features and interactive methods are used for selectively visualizing structures. Users can specify the structures and regions of interest on the rendered image and only visualize similar structures in the regions of interest. The proposed method introduces the parameters α and β, which are freely tuned by the user. The sensitivity α determines how easy it is to change visualization results by similarity, whereas the radius β determines the size of the region of interest.
By applying the proposed method to an image of a mouse brain acquired by two-photon microscopy, we visualize white matter, dendrites, and apical dendrites in which we verify the proposed method in an experiment in the region containing white matter. The proposed method is characterized by multidimensional features and interactive methods that are effective for visualizing target structures. Furthermore, we investigate the effect of parameters α and β on the results of visualization, where we find that a large β is suitable for search purposes and a smaller β is effective for small modifications of the visualization results. Because the suitable values of α and β depend on the situation, the proposed method promises that visualization can be implemented with flexibility, and that the parameters can be determined by the user to suit their particular visualization requirements.
Availability of supporting data
The data sets supporting the results of this article are included within the article.
This work was supported by Platform for Dynamic Approaches to Living System from the Ministry of Education, Culture, Sports, Science and Technology, Japan. This research was also supported by a Japan Society for the Promotion of Science (JSPS), Grant-in-Aid for Scientific Research for Young Scientists, Number 24680059.
- Matsuzaki M, Honkura N, Ellis-Davies GC, Kasai H: Structural basis of long-term potentiation in single dendritic spines. Nature. 2004, 429: 761-766. 10.1038/nature02617.View ArticlePubMed CentralPubMedGoogle Scholar
- Nicholson DA, Trana R, Katz Y, Kath WL, Spruston N, Geinisman Y: Distance-dependent differences in synapse number and AMPA receptor expression in hippocampal CA1 pyramidal neurons. Neuron. 2006, 50: 431-442. 10.1016/j.neuron.2006.03.022.View ArticlePubMedGoogle Scholar
- Nakashiba T, Cushman JD, Pelkey KA, Renaudineau S, Buhl DL, McHugh TJ, Rodriguez Barrera V, Chittajallu R, Iwamoto KS, McBain CJ, Fanselow MS, Tonegawa S: Young dentate granule cells mediate pattern separation, whereas old granule cells facilitate pattern completion. Cell. 2012, 149: 188-201. 10.1016/j.cell.2012.01.046.View ArticlePubMed CentralPubMedGoogle Scholar
- Kawakami R, Sawada K, Sato A, Hibi T, Kozawa Y, Sato S Yokoyama H, Nemoto T: Visualizing hippocampal neurons with in vivo two-photon microscopy using a 1030 Nm picosecond pulse laser. Sci Rep 2013, 3.,Google Scholar
- Pan F, Gan WB: Two-photon imaging of dendritic spine development in the mouse cortex. Dev Neurobiology. 2008, 68: 771-778. 10.1002/dneu.20630.View ArticleGoogle Scholar
- Trachtenberg JT, Chen BE, Knott GW, Feng G, Sanes JR, Welker E, Svoboda K: Long-term in vivo imaging of experience-dependent synaptic plasticity in adult cortex. Nature. 2002, 420: 788-794. 10.1038/nature01273.View ArticlePubMedGoogle Scholar
- Peng H, Ruan Z, Long F, Simpson JH, Myers EW: V3D enables real-time 3D visualization and quantitative analysis of large-scale biological image data sets. Nature Biotec. 2010, 28 (4): 348-353. 10.1038/nbt.1612.View ArticleGoogle Scholar
- Peng H, Tang J, Xiao H, Bria A, Zhou J, Butler V, Zhou Z, Gonzalez-Bellido PT, Oh SW, Chen J, Mitra A, Tsien RW, Zeng H, Ascoli GA, Iannello G, Hawrylycz M, Myers E, Long F: Virtual finger boosts three-dimensional imaging and microsurgery as well as terabyte volume image visualization and analysis. Nat Commun 2014, 5. http://dx.doi.org/10.1038/ncomms5342,
- Wan Y, Otsuna H, Chien C, Hansen C: An interactive visualization tool for multi-channel confocal microscopy data in neurobiology research. IEEE Trans Visual Compute Graphics. 2009, 15 (6): 1489-1496. 10.1109/TVCG.2009.118.View ArticleGoogle Scholar
- Heinmann T, Meinzer HP: Statistical shape models for 3D medical image segmentation: a review. Med Image Analysis. 2009, 13 (4): 543-563. 10.1016/j.media.2009.05.004.View ArticleGoogle Scholar
- Guerrero R, Wolz R, Rueckert D: Laplacian eigenmaps manifold learning for landmark localization in brain MR images. Med Imaging Comput Comput-Assist Interven 2011, 566-573,Google Scholar
- Levoy M: Efficient ray-tracing of volume data. ACM Trans Graphics. 1990, 9 (3): 256-261. 10.1145/78964.78965.View ArticleGoogle Scholar
- Cabral B, Cam N, Foran J: Accelerated volume rendering and tomographic reconstruction using texture mapping hardware. Proc. of Volume Visualization Symposium 1994, 91-98Google Scholar
- Lum E, Ma K: Lighting transfer functions using gradient aligned sampling. IEEE Visualization 2004, 289-296Google Scholar
- Kniss J, Kindleman G, Hanse C: Multidimensional transfer functions for interactive volume rendering. IEEE Trans Visual Comp Graphics. 2002, 8 (3): 270-285. 10.1109/TVCG.2002.1021579.View ArticleGoogle Scholar
- Kindlemann G, Whitaker R, Tasdizen T, Moller T, Curvature-based transfer-function for direct volume rendering methods and applications. IEEE Visualization 2003, 513-520,Google Scholar
- Frani AF, Nissen WJ, Vicken KL, Vergerver MA: Multiscale vessel enhancement filtering. Med Imaging Compute Computer Assisted Intervention. 1998, 1496: 130-137.Google Scholar
- Sato Y, Westin C, Bhalerao A, Nakajima S, Shiraga N, Tamura S, Kikins R: Tissue classification based on 3d local intensity structures for volume rendering. IEEE Trans Visual Comp Graph. 2000, 6 (2): 160-180. 10.1109/2945.856997.View ArticleGoogle Scholar
- Shikata H, Kitaoka H, Sato Y, Johkou T: Quantitative evaluation of spatial distribution of line structure in the lung for computer-aided diagnosis of pulmonary nodules. Systems Comp Japan. 2003, 34 (9): 58-70. 10.1002/scj.10222.View ArticleGoogle Scholar
- Praßni J, Ropinski T, Mensmann J, Hinrichs K: Shape-based transfer functions for volume visualization. IEEE Pacific Visualization 2010, 9-16,Google Scholar
- Caban J, Rheingans P: Texture-based transfer functions for direct volume rendering. IEEE Visual Comp Graph. 2008, 14 (6): 1364-1371. 10.1109/TVCG.2008.169.View ArticleGoogle Scholar
- Sereda P, Bartrolı A, Serlie I, Gerritsen F: Visualization of boundaries in volumetric data sets using LH histograms. IEEE Visual Comp Graph. 2006, 12 (2): 208-218. 10.1109/TVCG.2006.39.View ArticleGoogle Scholar
- Haidacher M, Patel D, Bruckner S, Kanitser A, Groller E: Volume visualization based on statistical transfer-function spaces. Proc. IEEE Pacific Visualization 2010, 17-24Google Scholar
- Correa CD, Ma KL: Size-based transfer functions: a new volume exploration technique. IEEE Trans Visuali Comp Graph. 2008, 14 (6): 1387-Google Scholar
- Wu Y, Qu H: Interactive transfer function design based on editing direct volume rendered images. IEEE Trans Visual Comp Graph. 2007, 13 (5): 1027-1040. 10.1109/TVCG.2007.1051.View ArticleGoogle Scholar
- Zhou J, Takatsuka M: Automatic transfer function generation using contour tree controlled residue flow model and color harmonics. IEEE Trans Visual Comp Graph. 2009, 15 (6): 1481-1488. 10.1109/TVCG.2009.120.View ArticleGoogle Scholar
- Läthén G, Lindholm S, Lenz R, Borga M: Automatic tuning of spatially varying transfer functions for blood vessel visualization. IEEE Trans Visual Comp Graph. 2012, 18 (12): 2345-2354. 10.1109/TVCG.2012.203.View ArticleGoogle Scholar
- Julia N, Maurer M, Mueller K: A high-dimensional feature clustering approach to support knowledge-assisted visualization. Comp Graph. 2009, 33 (5): 607-715. 10.1016/j.cag.2009.06.006.View ArticleGoogle Scholar
- Maciejewski R, Woo I, Chen W, Ebert D: Structuring feature space: a non-parametric method for volumetric transfer function generation. IEEE Trans Visual Comp Graph. 2009, 15 (6): 1473-1480. 10.1109/TVCG.2009.185.View ArticleGoogle Scholar
- Chen M, Ebert D, Hagen H, Laramee RS, van Liere R, Ma K-L, Ribarsky W, Scheuermann G, Silver D: Data, information, and knowledge in visualization. IEEE Comp Graph and Anim. 2009, 29 (1): 12-19. 10.1109/MCG.2009.6.View ArticleGoogle Scholar
- Kubisch C, Glaer S, Neugebauer M, Preim B, Vessel visualization with volume rendering, Visualization in Medicine and Life Sciences II (Springer). Mathematics Visualiz 2012: 109-134,Google Scholar
- Imanishi K, Nakao M, Kioka M, Mori M, Yoshida M, Takahashi T, Minato K: Interactive bone drilling using a 2D pointing device to support microendoscopic discectomy planning, Int. J Comp Ass Radiology Surgery. 2010, 5 (5): 461-469. 10.1007/s11548-010-0413-z.View ArticleGoogle Scholar
- Nakao M, Kurebayashi K, Sugiura T, Sato T, Sawada K, Kawakami R, Nemoto T, Minato K, Matsuda T: Visualizing in vivo brain neural structures using volume rendered feature spaces. Comp in Biol Med. 2014, 53: 85-93. 10.1016/j.compbiomed.2014.07.007.View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.