Hostname: page-component-788cddb947-xdx58 Total loading time: 0 Render date: 2024-10-07T13:39:58.948Z Has data issue: false hasContentIssue false

Bright-field to fluorescence microscopy image translation for cell nuclei health quantification

Published online by Cambridge University Press:  15 June 2023

Ruixiong Wang*
Affiliation:
Visual Information Laboratory, University of Bristol, Bristol, United Kingdom
Daniel Butt
Affiliation:
School of Biochemistry, University of Bristol, Bristol, United Kingdom
Stephen Cross
Affiliation:
Wolfson Bioimaging Facility, University of Bristol, Bristol, United Kingdom
Paul Verkade
Affiliation:
School of Biochemistry, University of Bristol, Bristol, United Kingdom
Alin Achim
Affiliation:
Visual Information Laboratory, University of Bristol, Bristol, United Kingdom
*
Corresponding author: Ruixiong Wang; Email: [email protected]

Abstract

Microscopy is a widely used method in biological research to observe the morphology and structure of cells. Amongst the plethora of microscopy techniques, fluorescent labeling with dyes or antibodies is the most popular method for revealing specific cellular organelles. However, fluorescent labeling also introduces new challenges to cellular observation, as it increases the workload, and the process may result in nonspecific labeling. Recent advances in deep visual learning have shown that there are systematic relationships between fluorescent and bright-field images, thus facilitating image translation between the two. In this article, we propose the cross-attention conditional generative adversarial network (XAcGAN) model. It employs state-of-the-art GANs (GANs) to solve the image translation task. The model uses supervised learning and combines attention-based networks to explore spatial information during translation. In addition, we demonstrate the successful application of XAcGAN to infer the health state of translated nuclei from bright-field microscopy images. The results show that our approach achieves excellent performance both in terms of image translation and nuclei state inference.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press

Impact Statement

Fluorescent images reveal information that cannot be revealed from bright-field images. Fluorescent image acquisition is a well-established technique but also complex. It requires sample preparation and can be time-consuming and error-prone. We present an image translation method based on deep learning which allows for revealing concealed information in bright-field images. The method has a high working efficiency, and whose speed of translating a fluorescent image ( $ 512\times 512\;{\mathrm{px}}^2 $ ) from a bright-field microscopy image stack can be less than a minute using publicly available Python libraries (mainly PyTorch) on mid-range, consumer hardware. In addition, our model is also able to infer the “health” state of nuclei from bright-field microscopy images simultaneous to the image translation process. Our method presents an alternative method for fluorescent image acquisition in biological research, for instance, where fluorescent labeling is not possible.

1. Introduction

Microscopy is a fundamental method for visualizing and interpreting cellular structures and processes invisible to the eye. Since around 1280 AD, the field of optics has evolved to go beyond the limits of human vision, eventually leading to the creation of simple single-lens microscopes(Reference Croft 1 ). Scientistcs’ understanding of light, optics, and materials has improved throughout the centuries and have brought forth far more powerful imaging technologies such as confocal laser scanning, fluorescent, electron, X-ray, and even acoustic microscopy(Reference Croft 1 ). Some cellular micro-structures are concealed in transmitted-light images, such as those acquired via bright-field, differential interference contrast (DIC) and phase contrast (PC) imaging. Fluorescent labeling can compensate for this limitation by highlighting specific cellular structures and can allow for their real-time tracking. The principle of fluorescence can be summarized as follows: electrons absorb energy from a photon (such as provided by the light source/laser beam), electrons become excited and rise to a higher energy level, and finally, the electrons lose their energy (as visible light) and return to their original energy level(Reference Lakowicz 2 ). Materials that display fluorescent properties are classified as fluorophores. Fluorophores can consist of fluorescent chemicals (e.g., nuclear dyes such as DAPI and Hoechst) and fluorescent proteins (e.g., green fluorescent protein [GFP]) bound to antibodies(Reference Murphy 3 , Reference Jia, Zhu, Wu and Wu 4 ). Additionally, some endogenous biomolecules such as the coenzymes: nicotinamide adenine dinucleotide (NADH) and flavin adenine dinucleotide (FAD) display fluorescent properties(Reference Murphy 3 , Reference Andersson, Baechi, Hoechl and Richter 5 ). Fluorescent imaging is especially useful in the study of the nucleus or DNA as it is very difficult to observe DNA with traditional light microscopes. This can be remedied by staining it with Hoechst or DAPI, fluorescent chemicals specific in binding to DNA. The Hoechst stain is the preferred choice as it provides greater cell permeability and is less cytotoxic to cells compared to DAPI. It is excited by UV light (~350 nm) and emits blue light (~460 nm)(Reference Murphy 3 , Reference Bucevičius, Lukinavičius and Gerasimaitė 6 ). Therefore, Hoechst is incredibly useful in studying nuclear morphology during apoptosis.

Apoptosis, referred to as programmed cell death, is a mechanism within multi-cellular organisms that occurs during embryonic development (formation of body parts). It controls the cell population, that is, by removing aged or unhealthy cells. It is also extremely important in destroying cells that have become infected or damaged by harmful agents, to prevent them from influencing the rest of the organism(Reference Elmore 7 ). The morphological features of apoptosis entail cell and nuclei shrinkage (pyknosis), followed by the breakdown of nuclear material into smaller fragments (karyorrhexis). Finally, the organelles, DNA fragments, and cytoplasm are deposited into smaller membrane (apoptotic) bodies which break away from the main cell body (budding or blebbing) and are destroyed by the host’s immune cells(Reference Elmore 7 ). Whilst fluorescence imaging is a powerful technique, the disadvantages of following these events in live cells are also prominent. Fluorescent nuclear stains including Hoechst, bind to the DNA and as a consequence prevent DNA replication during cell division, essentially ending the viability of the cells(Reference Nikolai 8 ). Phototoxicity is also increased during live cell fluorescence imaging. As the cells are exposed to high intensities of light, the fluorescent labels produce reactive oxygen species in their excited states, which in turn damages the structure and functionality of critical biomolecules(Reference Laissue, Alghamdi, Tomancak, Reynaud and Shroff 9 ). Additionally, fluorophores can eventually become irreversibly damaged if they are not imaged correctly or for an extended period, referred to as photobleaching. It has been recommended to employ the optimum imaging parameters, spend minimum time searching for perfect images, and remove the oxygen from the medium. Regardless, all of these factors are still genuine limitations experienced by users of fluorescence microscopy(Reference Waters, Sluder and Wolf 10 ). Studies on apoptosis generally start with healthy cells imaged under optimal conditions. Such conditions need to be established first and could already introduce cell damage. The onset of the apoptotic process can then be induced by the addition of drugs or other stressors. Together with the time needed to prepare the samples (including repeats), which can be time-consuming, the whole process of fluorescence imaging can lead to the introduction of artifacts and ultimately failure of the experiment. To overcome these drawbacks, researchers have started to explore the latent information hidden in the bright-field images to simulate the fluorescent images using computer vision technologies. Up to now, there have been very few conventional methods for fluorescent image translation, the reason being that the amount of latent features within images is too big to be extracted manually, and it is even harder to align features between the two image domains.

The rapid evolution of machine-learning techniques provides new opportunities for applying image translation in this area. The first image translation method was implemented by Christiansen et al. in 2018(Reference Christiansen, Yang and Ando 11 ). They established a model based on deep learning that reliably predicts labels from transmitted-light images. Deep learning is a machine-learning technique that specifically uses deep neural networks(Reference LeCun, Bengio and Hinton 12 , Reference Goodfellow, Bengio and Courville 13 ). In the field of image processing, it was first adopted for image classification and segmentation(Reference Rawat and Wang 14 , Reference Moen, Bannon, Kudo, Graf, Covert and van Valen 15 ). Deep convolutional neural networks (CNNs) are employed to extract useful and unapparent information from images automatically, circumventing the problems of manual feature extraction. The work of Christiansen et al. determined that there were strong relationships between bright-field images and the corresponding fluorescent images(Reference Christiansen, Yang and Ando 11 ). The idea of translating a bright-field image into a fluorescent image using deep learning is thus feasible.

The generative adversarial network (GAN) is one of the most popular neural networks and has been widely used in image translation tasks(Reference Goodfellow, Pouget-Abadie and Mirza 16 Reference Taigman, Polyak and Wolf 19 ). It has the advantage of solving the blurring problem that generally occurs when generating synthetic images(Reference Isola, Zhu, Zhou and Efros 20 ). GANs were recently introduced in fluorescent image translation and showed significant advantages in generating multiple channels of fluorescent images simultaneously and predicting sub-cellular structure details(Reference Li, Zhang and Qiao 21 Reference Lee, Oh, Kang, Her, Kim and Jeong 24 ). In some existing work, each channel of the output represents a specific labeling result with a different stain or dye derived from a single bright-field image(Reference Li, Zhang and Qiao 21 Reference Nguyen, Bui and Thai 23 , Reference Zhang, Ning and Hui 25 ). For nuclei belonging to the same species but in different states, such as the instance shown in Figure 1, previous approaches would have difficulties in distinguishing between the different states of the nuclei, one such example of this is shown in Section 5. In this work, we propose a novel deep-learning model based on a conditional GAN (cGAN) aimed at translating bright-field images to fluorescent images with simultaneous nuclei health state inference. In addition, we extracted the spatial dependency information from the state inference process and used it as supplementary information for image generation. Inspired by self-attention mechanisms(Reference Vaswani, Shazeer, Parmar, Guyon, Von Luxburg and Bengio 26 , Reference Zhang, Goodfellow, Metaxas and Odena 27 ), which have been used for calculating long-distance spatial dependencies during image generation, attention-based modules were included in our model for learning spatial dependencies. The attention-based modules take feature maps from both image generation and segmentation paths and feed them into the classification mechanism, a process referred to as cross-attention.

Figure 1. Information comparison between bright-field light microscopy (left) with fluorescence light microscopy (right) of the same cells. Green boxes indicate zoomed areas of healthy nuclei and magenta boxes indicate apoptotic nuclei with yellow arrows pointing to the position of the cells in the overview images.

Following this general introduction to the subject, in Section 2, we set the scene by introducing previous work on fluorescent image translation using deep learning and in particular, approaches based on GANs. We also describe the attention network mechanism. In Section 3, we elaborate on the architecture of our model, including specific details of the generator and discriminator. Section 4 introduces the methodology, including the biological experiment, microscopy image acquisition, image pre-processing, and other preparation work for model training. Section 5 presents the experimental results. Section 6 is the ablation study. Finally, Section 7 concludes the article.

2. Theoretical Preliminaries

2.1. Image translation using convolutional neural network

In silico labeling (ISL) was one of the first published deep-learning approaches that could determine the correlation between transmitted-light microscopy images and fluorescent images(Reference Brent and Boucheron 28 ). It succeeded in predicting the location and intensity of nuclear labeling on bright-field images with good accuracy. The fluorescent images generated by the ISL model could be used in quantifying the dimensions of cellular structure and classifying the physiological state of cells(Reference Christiansen, Yang and Ando 11 ). Specifically, the ISL model performed the task of pixel-to-pixel classification and Christiansen et al. used a z-stack image, consisting of z slices of bright-field images, as the input for multi-channel fluorescent image prediction. The model was inspired by the popular U-Net(Reference Ronneberger, Fischer and Brox 29 ) and inception networks(Reference Szegedy, Liu and Jia 30 ). For each translation, five image patches were cropped out of the input bright-field image stack with different side lengths at the same centroid location and simultaneously fed into five independent input computational heads of the model, respectively; the outputs from the five computational heads were concatenated together and transmitted to the final prediction computational head of model to synthesize the fluorescent images. As each input computational head processed one side-length area, it allowed the model to learn spatial information from multiple spatial extents. The other special point of this network was that it was composed of repeated modules; the mutually independent modules had the same architecture but different parameters that controlled the number of hidden convolutional kernels. The ISL model implemented the translation piece by piece. A stack of bright-field image patches with $ 250\times 250\;{\mathrm{px}}^2 $ was fed into the model. The ISL model predicted an 8-bit gray-scale image, where each unique gray-scale value had a probability. For each pixel, these probabilities were represented by a 256-element vector. Such a process ensured accuracy; however, it sacrificed computational efficiency, as the model was large and the parameters for each repeated module were hard to tune(Reference Christiansen, Yang and Ando 11 ).

Soon after the first trial for fluorescent image translation, Ounkomol et al. (Reference Ounkomol, Seshamani, Maleckar, Collman and Johnson 22 ) proposed their work for label-free prediction of multi-channel immuno-fluorescence (IF) images from 3D transmitted-light or 2D electron microscopy (EM) images. As EM images contain plenty of sub-cellular structure information, this model was capable of presenting more intracellular details. The model can also automatically register EM images on large target IF images. The model for label-free prediction was a CNN based on the U-Net architecture. It contained down- and up-sampling substructures with shape concatenation processes. Compared to an ISL model, its architecture was much more intuitional and close to the initial U-Net architecture(Reference Ounkomol, Seshamani, Maleckar, Collman and Johnson 22 ).

2.2. Image translation based on generative adversarial neural network

Image translation approaches with deep learning have gone through a breathtaking evolution in recent years. The initial attempt using CNNs was an impressive achievement, but its results tended to be blurry. This was because the loss function was minimized by averaging all plausible outputs(Reference Isola, Zhu, Zhou and Efros 20 ). To overcome the disadvantages of the blurring issue, the generative adversarial neural network (GAN) was proposed(Reference Goodfellow, Pouget-Abadie and Mirza 16 , Reference Denton, Chintala, Fergus, Cortes and Lawrence 31 Reference Salimans, Goodfellow, Zaremba, Cheung, Radford, Chen, Lee and Sugiyama 33 ). The basic concept for GAN was adding another CNN to the generative model for discriminating whether the output is real or fake. This adversary system was trained together and encouraged the generative network to produce more convincing results and prevent the outputs from being blurry.

The original GANs generated outputs from random vectors(Reference Gui, Sun, Wen, Tao and Ye 34 , Reference Yi, Walia and Babyn 35 ); however, subsequent variants have since been derived. For example, Li et al. employed CycleGAN(Reference Zhu, Park, Isola and Efros 18 ) for unsupervised content-preserving transformation for optical microscopy (UTOM)(Reference Li, Zhang and Qiao 21 ). CycleGAN employed two generators to transform images between two domains back and forth, respectively, and two discriminators were used for penalizing whether the generated images belong to each domain; images transformed forward and then transformed backward should be the same. One speciality of CycleGAN was that images from the two domains did not need to be corresponding, so it allowed for unsupervised translation(Reference Zhu, Park, Isola and Efros 18 ). Meanwhile, the UTOM model used a saliency constraint to maintain the saliency within the images unchanged during transformation to avoid distortion of content. Generators in UTOM had the same number of input and output channels to simplify the transactions. It was shown that UTOM achieved comparable performance to the supervised CNN model without having any paired training datasets. However, the training dataset for UTOM was larger than the supervised model, otherwise, the accuracy of translation would be limited(Reference Li, Zhang and Qiao 21 ).

UTOM achieved a great performance for image translation, but the limitation of the dataset could not be neglected. Therefore, a model that combined the advantages of supervised learning and GANs could be another solution. cGAN was a promising approach suitable for tasks with limited but paired datasets(Reference Mirza and Osindero 36 ). The discriminator in cGANs not only judged whether the outputs were real or fake but also determined whether the inputs and outputs were consistent. As a supervised learning process, the loss of the generator was a combination of the discriminating result and the distance between the outputs and ground truth. Isola et al. (Reference Isola, Zhu, Zhou and Efros 20 ) proposed an image-to-image translation model with cGAN. The model succeeded in image generation from sparse annotations, future frame prediction, or photo reconstruction. Their generator for cGANs was inspired by U-Net(Reference Ronneberger, Fischer and Brox 29 ), and the discriminator used a specially designed “PatchGAN” classifier to enhance the spatial penalization(Reference Isola, Zhu, Zhou and Efros 20 , Reference Li and Wand 37 ). This approach was widely used as the benchmark for style transfer of cellular images(Reference Hollandi, Szkalisity and Toth 38 ).

Based on the concept of cGAN, Han et al. (Reference Han and Yin 39 ) produced a model for cellular image translation. The model was established for the DIC to PC transition. The only difference in the architecture from the original cGAN model was the model took the predefined cell masks as additional inputs and the system consisted of two discriminators, one for DIC–PC pairs and one for mask–PC pairs. This model showed that the generator performed better compared to only DIC images for inputs.

After the success of applying cGANs in microscopy image translation, Nguyen et al. (Reference Nguyen, Bui and Thai 23 ) designed a complex system with multiple cGANs applied on microscopy images taken from breast-cancer and bone-osteosarcoma cell lines. The system implemented not only the transformation of images from fluorescent images to PC images but also the interpretation between different fluorescent channels. The architectures for each translation model were basic pixel-to-pixel image translation networks introduced in the work of Isola et al. (Reference Isola, Zhu, Zhou and Efros 20 ), and they implemented a novel algorithm for evaluating the generated result. Their work achieved great success in the prediction of subcellular localization of proteins with other proteins on well-registered images. It proved that a cellular image translator based on cGAN was an effective tool for studies about relationships between proteins and organelles and to visualize the subcellular structure(Reference Nguyen, Bui and Thai 23 ).

Lee et al. provided the DeepHCS++ model for the fluorescent image translation task(Reference Lee, Oh, Her and Jeong 40 ). The architecture of the DeepHCS++ model consisted of two parts: transformation and refinement networks. The transformation network was similar to the label-free model of Ounkomol et al.; however, instead of using one up-sampling decoder for multi-task learning, three independent decoders were applied to produce three channels of fluorescent images, respectively. The intermediately produced fluorescent images were then fed into three U-Net shape refinement networks for the final predictions. Conditional adversarial loss calculations were only applied to the final output images rather than intermediate products. Results showed that the performance for translation-refinement networks was better than single translation networks(Reference Lee, Oh, Her and Jeong 40 ).

2.3. Attention neural network

Images generated by GANs derived from lower-resolution feature maps and higher-level details are calculated with spatially local points. This process has a significant limitation as it only takes spatially local information for calculations(Reference Zhang, Goodfellow, Metaxas and Odena 27 , Reference Cordonnier 41 ). However, the ISL model showed that the translation required information from multiple spatial extents of correlated bright-field images, which is why it took five scales of inputs. Fortunately, self-attention GANs (SAGANs) provided an approach for long-range dependency modeling for image generation.

Self-attention mechanisms allowed for computing dependencies within a single sequence without the restriction of position(Reference Vaswani, Shazeer, Parmar, Guyon, Von Luxburg and Bengio 26 , Reference Cheng, Dong and Lapata 42 , Reference Parmar, Vaswani and Uszkoreit 43 ). It had been successfully used in sequence-to-sequence processing such as reading comprehension, textual analysis and language translation. This approach can be transferred to spatial analysis in image processing(Reference Guo, Xu and Liu 44 ). Zhang et al. proposed the SAGAN, which aimed to calculate the dependencies of one position on all others through feature maps. The result showed that the SAGAN was an effective model for the calculation of long-range dependencies and upgraded the qualities of output images(Reference Zhang, Goodfellow, Metaxas and Odena 27 ). Besides, SAGANs applied spectral normalization(Reference Miyato, Kataoka, Koyama and Yoshida 45 Reference Odena, Buckman and Olsson 47 ) to the generator and discriminator to enhance the stability of GAN training. The mechanisms of SAGAN were introduced in our model, further details will be elaborated in the next section.

3. Proposed Approach

The framework proposed in the translation task relies on a cGAN. Similar to the basic GAN network, it also includes a conditional generator system for image translation, in conjunction with a discriminator to evaluate its outputs. Both generator and discriminator are composed of a series of sub-modules. What is different is that the generator contains two generation paths: one for image translation and one for nuclei segmentation. Thereinto, nuclei segmentation is implemented by producing masks for nuclei with different health states. Attention modules in the network are used for transferring information between the two generating paths within the generator. The module is based on self-attention mechanisms, and we refer to it as the cross-attention module. Because of the dual-output generator, the loss function consists of a weighted combination of both paths.

3.1. Generator

The architecture of the generator is inspired by the U-Net(Reference Ronneberger, Fischer and Brox 29 ) and ResNet(Reference He, Zhang, Ren and Sun 48 ) networks. U-Net was the first network designed for biological image segmentation. It contains an encoder–decoder system linked by skip connections. The encoder consists of a down-sampling process to transform the images to feature maps with more channels but smaller sizes; on the contrary, the decoder is an up-sampling process that transforms the intermediate feature maps to the output size. Intermediate outputs from the encoder are concatenated to the decoder which enhances spatial accuracy. One difference in our model is its two independent generation paths for image translation and nuclei semantic segmentation. Each channel of the image translation output represents one health state of the nuclei, meanwhile, the nuclei segmentation path provides binary masks for each state or background. The generator is constituted by a series of sub-modules, and each sub-module corresponds to one down-sampling or up-sampling process. Additionally, there is one bottleneck module to connect the encoder and decoders. Figure 2 shows the structure of the sub-modules. Skip connections for residual learning are employed in the sub-modules to enhance the robustness of the model. Each sub-module contains three parts, a shape-invariant convolutional layer, a sampling convolutional layer, and a residual connection(Reference He, Zhang, Ren and Sun 48 ). The final activation function layer for the image generation decoder is Tanh to restrict the range of image values between −1 and 1. The output of the mask generation process is sent into a SoftMax function to calculate the probability of pixels belonging to each category. The output of the mask generation path has one more channel than the image generation path, which is for the background probability prediction. Finally, the image generation outputs are multiplied by the probability results from the segmentation path to produce the translation result. Details of the architecture of the generator are shown in Figure 3.

Figure 2. Network architectures of sub-modules for generator. (a) Down-sampling sub-module. (b) Bottleneck sub-module. (c) Up-sampling sub-module.

Figure 3. Generator architecture. The numbers in each sub-module indicate the numbers of input and output channels, respectively. Architectures of sub-modules are presented in Figure 2, and attention modules are shown in Figure 4.

3.2. Attention-based module

Attention algorithm-based modules are utilized to learn long-distance spatial dependencies. They are inserted between the decoders to exchange spatial information during image generation. The initial attention algorithm module is self-attention which calculates the spatial dependencies within the feature maps themselves. However, we want the attention module to also share the spatial dependencies between two generation paths. As our attention module derives from the self-attention module, we will describe the structure of the self-attention module first, and then compare the improvement of our attention-based module. The self-attention network is presented in Figure 4a, and the processing steps are listed below.

  1. 1. Send the input feature map stack into three 1 × 1 convolutional layers and reshape the outputs into 1D sequences, $ q $ , $ k $ , and $ v $ .

  2. 2. Calculate the attention of feature $ q $ and $ k $ .

  3. 3. Feed the attention output into a SoftMax activation function.

  4. 4. Calculate the attention of the SoftMax output with feature $ v $ .

  5. 5. Reshape the 1D sequence back to the original shape.

  6. 6. Send the feature into another 1 × 1 convolutional layer to keep the self-attention module output the same shape as the input.

  7. 7. Add the self-attention output back to the feature map stack and multiply with a learnable parameter.

Figure 4. Attention module architectures. (a) Architecture of the self-attention module. (b) Architecture of the cross-attention module.

In our model, our aim is to share the spatial dependency information from both generation paths. Therefore, we extract hidden features from both generation paths using in Step 1, ( $ {q}_{image} $ , $ {k}_{image} $ , and $ {v}_{image} $ from image feature domain; $ {q}_{mask} $ , $ {k}_{mask} $ , and $ {v}_{mask} $ from mask feature domain). Our attention module contains two self-attention calculation paths which remain the same as the self-attention module; at the same time, hidden feature $ {q}_{image} $ (from the image path), $ {k}_{mask} $ and $ {v}_{mask} $ (from the mask path) are sent to the algorithm from Step 2 to Step 6 to measure the correlation between decoders. This process is similar to the work in by Hou et al. (Reference Hou 49 ). These are concatenated with the image self-attention outputs and sent to the next layer in the image generation path. As previous work proves that a U-Net with one encoder–decoder pair is enough for biomedical image semantic segmentation(Reference Ronneberger, Fischer and Brox 29 ), we decided not to make the mask generation process as complex as image generation. Therefore, the correlated attention algorithm is not applied to the mask generation path. The module is named cross-attention as it takes inputs from both paths and feeds them back. The cross-attention modules are inserted between feature map layers of the same size from different generation decoders. The architecture of the cross-attention module is presented in Figure 4b.

3.3. Discriminator and spectral normalization

The discriminator does not produce translation results but instead guides the training destination of the generator and self-upgrades simultaneously. Normally, the discriminator is a down-sample process which assesses whether the input is good or not. Isola et al. introduced a discriminator named PatchGAN for their image-to-image translation network(Reference Isola, Zhu, Zhou and Efros 20 ). PatchGAN outputs N × N patches and penalizes the patches independently, where the size of patches can be much smaller than the full size of the image. They found that the PatchGAN in the discriminator performed well in enforcing correctness at high frequencies and that low-frequency correctness could be achieved using conventional methods, such as the L1 distance restrictor. The discriminator takes the concatenation of bright-field image stacks and generated translated images as the input and runs a series of convolutional calculations. The architecture of the discriminator is shown in Figure 5; it is also made up of down-sampling sub-modules which have a similar structure as the one used in the encoders of the generator. Self-attention modules are applied in our discriminator; they are inserted after the down-sampling sub-modules at higher levels.

Figure 5. Discriminator architecture. The numbers in each sub-module indicate the numbers of input and output channels, respectively. Numbers in attention boxes represent the size of input feature maps.

To improve the performance of the discriminator, Miyato et al. created a new normalization method called spectral normalization and applied it to the discriminator of their GAN(Reference Miyato, Kataoka, Koyama and Yoshida 45 ). They discovered that spectral normalization can stabilize the training of the discriminator, as it tuned fewer extra hyper-parameters and had lower computational costs compared to other regularization techniques. In the work of Zhang et al., spectral normalization was applied to both the generator and discriminator. This achieved better performance compared to applying it only to the discriminator. Hence, we also apply spectral normalization to our discriminator and generator(Reference Zhang, Goodfellow, Metaxas and Odena 27 ).

3.4. Loss function and evaluation metrics

The fundamental working mechanism of GANs can be understood as a competition between the generator and discriminator. The generator aims to produce outputs that fool the discriminator into being unable to distinguish whether its products are real or fake. On the contrary, the discriminator aims to train itself to not be fooled by the generator. Mathematically, this process can be regarded as finding the solution to minimum and maximum objective functions. In cGAN, the generation process is under given conditions, in the translation task the output fluorescent images need to be correlated to the input bright-field image stacks. Therefore the discriminator needs to take the input bright-field image stacks as the condition during the penalization. The objective function of cGAN is shown in Equation (1):

(1) $$ {\mathcal{L}}_{cGAN}\left(G,D\right)={\unicode{x1D53C}}_{x,y\sim {P}_{data}\left(x,y\right)}\left[\log D\left(x,y\right)\right]+{\unicode{x1D53C}}_{x\sim {P}_{data}(x)}\left[1-\log D\left(x,G(x)\right)\right], $$

where $ {P}_{data} $ is the ground truth dataset, $ x $ is the input bright-field image, and $ y $ is the target fluorescent image; $ G $ and $ D $ represent the generator and the discriminator, respectively. The generator is trained to minimize the objective function and the discriminator is trained to maximize it; the generator is described in Equation (2):

(2) $$ {G}^{\ast }=\arg \underset{G}{\min}\underset{D}{\max}\left({\mathcal{L}}_{cGAN}\left(G,D\right)+\lambda {\mathcal{L}}_{Image}(G)\right), $$

where $ {\mathcal{L}}_{Image} $ is the conventional loss between the generated outputs and the ground truth targets, such as L1 loss (details to follow). $ \lambda $ is the weight that balances the two types of losses.

The total loss for the system is a combination of adversarial loss and conventional losses. Typical conventional losses include mean absolute error loss (MAE) or mean squared error (MSE) loss, also known as L1 and L2 distances, respectively. The conventional loss function does well in penalizing the low-frequency errors which are supplementary to adversarial loss. In previously published approaches(Reference Isola, Zhu, Zhou and Efros 20 , Reference Lee, Oh, Her and Jeong 40 ), the weight of the conventional loss is normally one or two orders of magnitude higher than the adversarial loss in cGAN, we will follow this precedent. Meanwhile, MAE losses are used rather than MSE losses in image generation as research showed that MAE-based methods cause less blurring in practice(Reference Isola, Zhu, Zhou and Efros 20 ). Moreover, based on the experience from Hore et al., we introduce a structural similarity index measure (SSIM) distance for the conventional loss calculation(Reference Lee, Oh, Her and Jeong 40 ). SSIM is a method for measuring the similarity between two images; it is a perception-based model that considers the degradation in structural information, which is expressed in Equation (3):

(3) $$ SSIM\left(x,y\right)=\frac{\left(2{\mu}_x\hskip0.35em {\mu}_y+{c}_1\right)\left(2{\delta}_{xy}+{c}_2\right)}{\left({\mu}_x^2+{\mu}_y^2+{c}_1\right)\left({\delta}_x^2+{\delta}_y^2+{c}_2\right)}, $$

where $ {\mu}_{\left(\cdot \right)} $ and $ {\delta}_{\left(\cdot \right)} $ are the mean and variance of $ x $ and $ y $ , $ {\delta}_{xy} $ is the covariance of $ x $ and $ y $ ; $ {c}_1 $ and $ {c}_2 $ are constant values based on the intensity range of $ x $ and $ y $ . The output value of the SSIM function is from 0 to 1, the more similar they are the higher the value is, and hence, the SSIM distance of $ x $ and $ y $ can be calculated as $ {\mathcal{L}}_{SSIM}=1- SSIM\left(y,G(x)\right) $ . Thus, the conventional loss for image generation is presented in Equation (4):

(4) $$ {\mathcal{L}}_{Image}=\left(1-\alpha \right){\mathcal{L}}_{L1}\left(y,G(x)\right)+\alpha {\mathcal{L}}_{SSIM}\left(y,G(x)\right), $$

where $ \alpha $ is the weight to balance L1 loss and the SSIM distance, $ \alpha \hskip0.35em \in \hskip0.35em \left[0,1\right] $ .

Other than losses from the image generation path, the output of the mask generation path predicts the possibilities of pixels belonging to each category. The loss of segmentation from the mask generation is defined as the classical cross-entropy. The final loss for the generator is a weighted sum of image generation losses, including conventional loss and adversarial, and mask loss, total loss calculation is shown in Equation (5):

(5) $$ {\mathcal{L}}_{Total}={\mathcal{L}}_{cGAN}+{\mu}_1{\mathcal{L}}_{Image}+{\mu}_2{\mathcal{L}}_{Mask}, $$

where, $ {\mu}_1 $ and $ {\mu}_2 $ are weights for each loss. In practice, we chose a dynamic weight for $ {\mu}_2 $ , the loss for mask generation. A higher value for $ {\mu}_2 $ was applied to ensure the mask generation reaches the convergence point first so that the mask generation path provides highly credible spatial information to the cross-attention module, and gradually lowers the weight of the mask to enhance the image generation.

In addition, the performance of the model is evaluated using three metrics: the L1 distance (i.e., the MAE), the structural similarity index (SSIM), and the standard peak signal-to-noise ratio (PSNR)(Reference Hou, Chang, Ma, Shan, Chen, Wallach and Larochelle 50 ).

4. Experimental Setup

4.1. Biological experiment and microscopy operation

In this section, we further describe the methodological approach from a biological point of view, providing further motivation for the work and presenting the datasets employed.

The cells used in this study are Chinese hamster ovary (CHO)-K1 cells [ATCC-CCL-61], which are induced to undergo apoptosis, to analyze and predict the state of their nuclei. These are cultured in Ham’s F-12K (Kaighn’s) Medium (Life Technologies Limited, Gibco, Paisley, Scotland, United Kingdom) supplemented with 10% fetal bovine serum (FBS). They are grown in T-75 flasks (10 ml of culture medium) and incubated at 37°C and 5% CO2.

  1. 1. When the cells have reached minimum confluency of 60%, ~100,000 cells are seeded onto glass coverslips (13 mm diameter) in 500 μL of medium and left to adjust overnight. All rinsing steps are done with Phosphate buffered saline (PBS).

  2. 2. To induce apoptosis, the cells are exposed to 500 μL of 1 μM of staurosporine (in medium, Apexbio Technology LLC, Houston, Texas, United States) and incubated for 0, 1, 3, 6, and 8 hr.

  3. 3. Once the respective time has been reached, the cells are fixed in 4% Paraformaldehyde (PFA) in PBS for 30 min.

  4. 4. The nuclei are stained with Hoechst solution (1:10,000 dilution in medium, Thermofisher Scientific, Pierce Biotechnology, Rockford, Illinois, United States) for 10 min.

  5. 5. Finally, the coverslips are mounted onto glass slides and left to curate overnight.

The cells are imaged on a Leica SP5 confocal laser scanning microscope using an HC PL APO CS2 63×, 1.4 NA oil objective. Hoechst-stained nuclei are excited with a 405 (nm) diode laser (laser power: 7%). Single-plane bright-field images and fluorescent image stacks are taken of the nuclei. The fluorescent image stacks have a z-axis range of 7.2 μm and consist of 24 slices spaced 0.3 μm apart.

4.2. Dataset preparation

Our method is designed to produce fluorescent images together with semantic segmentation. The most important part of the image pre-processing pipeline was segmenting the fluorescent images based on the health state of each nucleus. The states of the nuclei change over time, that is, after 3 hr nuclei start to split. In the pre-processing phase, fragmented nuclei were merged using Gaussian and median filters. Subsequently, a marker-controlled watershed segmentation process was applied to separate adherent nuclei. All steps in the pre-processing pipeline are listed below and shown in Figure 6.

  1. 1. Resize raw images to an appropriate size that maintains adequate information. In this work, we resize the images from $ 512\times 512\;{\mathrm{px}}^2 $ to $ 256\times 256\;{\mathrm{px}}^2 $ .

  2. 2. Apply min–max normalization for each raw fluorescent image and rescale to 0 to 255 (8-bit intensity range).

  3. 3. Concatenate fluorescent images along z-stack. Choose the maximum value along the z-stack for each pixel and generate the maximum image.

  4. 4. Apply contrast limited adaptive histogram equalization (CLAHE)(Reference Pizer, Johnston, Ericksen, Yankaskas and Muller 51 ) to each image. Set the contrast limiting value to be 2, and the size of tiles to be $ 32\times 32\;{\mathrm{px}}^2 $ .

  5. 5. Apply Gaussian smoothing to images using a $ 3\times 3\;{\mathrm{px}}^2 $ kernel. In practice, we apply it four times so that fragmented spots from the same nuclei are connected.

  6. 6. Apply median filtering with a $ 5\times 5\;{\mathrm{px}}^2 $ kernel. In practice, we apply it 12 times to remove salt-and-pepper noise from the images.

  7. 7. Calculate the Otsu threshold for the images and generate the mask for nuclei.

  8. 8. Calculate the distance for pixels within the mask to the mask boundaries. Then select the local maximum points on the distance maps and label them as the kernel for individuals.

  9. 9. Apply the watershed algorithm(Reference Beucher and Lantuejoul 52 , Reference Beucher 53 ) to segment the masks.

  10. 10. For nuclei incubated in staurosporine for less than 8 hr, the number of apoptotic cells is much less than healthy cells. The split nuclei can be easily filtered out manually.

  11. 11. For nuclei at 8 hr. Apply individual segmented nucleus masks to the maximum fluorescent images. Calculate the standard deviation of the intensity value within each segmented nucleus image. Use the Otsu thresholding method(Reference Otsu 54 ) to separate these nuclei into two groups. The group which has a higher value of standard deviation corresponds to fragmented nuclei and to intact nuclei. Inspect the output of the segmentation process based on standard deviation, and manually correct the splitting results.

  12. 12. Individual masks belonging to the same group are added together and used to generate the masks for healthy nuclei and apoptotic nuclei.

  13. 13. Apply both nuclei masks on the center layer of the z-stack fluorescent images after CLAHE. Each fluorescent image is divided into two channels. The two-channel fluorescent images are used as the ground-truth targets for training.

  14. 14. For bright-field images, apply min–max normalization for rescaling and concatenation along z-stack.

Figure 6. Dataset preparation process. (a) Maximum intensity along z-stack of fluorescent images, (b) threshold and watershed segmentation output, (c) automatic classification result, (d) manually revised result, the revised individual in the yellow circle, (e) example of image dataset for training, contains bright-field images, split fluorescent images, and masks.

4.3. Dataset augmentation

Due to the limited number of images in our dataset, data augmentation was applied for the training of the model. Random flips and rotations were applied, followed by a random crop of $ 128\times 128\;{\mathrm{px}}^2 $ out of $ 256\times 256\;{\mathrm{px}}^2 $ . Random flips and rotations encourage the model to not be restricted by the orientation of the input images. A random crop of images helps individual nuclei maintain the same sizes and ratios (width to height) compared to the original images, but the locations of nuclei on images vary at each iteration. Data augmentation ensures the inputs to the model are different for every epoch and prevents over-fitting of the model.

4.4. Training details

The code was implemented using Python and PyTorch. This work was carried out using the BlueCrystal Phase 4 facility of the Advanced Computing Research Centre, University of Bristol (http://www.bristol.ac.uk/acrc/). The system is equipped with Nvidia P100 GPU with 16 GB of RAM. The optimizer for the generator and discriminator used here was the Adam optimizer with beta values of 0.5 to 0.999(Reference Kingma and Ba 55 ), while the learning rates were $ {10}^{-5} $ and $ {10}^{-4} $ , respectively(Reference Heusel, Ramsauer, Unterthiner, Nessler, Hochreiter, Guyon and Von LuxBurg 56 ). The batch size for training was 8. The total number of learning epochs was 4,500. Cross-validation was applied to the training dataset, with the number of images for training, validation, and testing being 66:12:8. The training and validation datasets were rearranged every 50 epochs. As mentioned above, the weight for the mask generation loss varies during each iteration; the initial weight was set to 250 and decayed 10% every 1,500 epochs and no lower than 2.5.

5. Results

In our model, whose architecture is shown in Figure 3, there are five up-sampling layers. Therefore, there are four intervals between adjacent up-sampling layers for inserting self-attention modules, so theoretically, there should be $ {2}^4-1 $ models to be tested. We used four-digit binary numbers to indicate the locations of the attention modules. Zhang et al. found the attention level applied at feature maps with larger sizes received better performance(Reference Zhang, Goodfellow, Metaxas and Odena 27 ). Therefore, in our strategy of choosing the insertion sites, we filled in lower intervals (where feature maps had larger sizes). As such, we selected three models for performance evaluation and comparison: 0001 (attention module only appeared at the last interval), 0011 (last two intervals), and 0111 (last three intervals). We also selected two models as references. The first was the cross-attention cGAN (XAcGAN) model, but with no cross-attention module applied; this was labeled as 0000. The second was the original image translation model using cGAN, the pixel-to-pixel image translation model, introduced in Isola et al.’s work(Reference Isola, Zhu, Zhou and Efros 20 ). The second reference model had six down-sampling layers and six up-sampling layers with skip connections. This second reference model had no residual network module or spectral normalization in either the generator or discriminator.

The prediction results for XAcGAN models are listed in Table 1 together with the pixel-to-pixel model. As expected, the XAcGAN models’ performances were significantly better than the pixel-to-pixel model in all evaluation metrics. Of the three tested models, model 0011 received the best scores, which are highlighted in bold. Figure 7 shows the prediction from the XAcGAN 0011 model.

Table 1. Performance of cross-attention conditional GAN model.

Abbreviations: PSNR, peak signal-to-noise ratio; SSIM, structural similarity index measure.

a The inserted location is presented by a four-digit binary number, the first digit represents the position between the first and second down-sampling layers and can deduce the rest like this. Digit “1” means an attention-based module is inserted at this position, and vice versa.

Figure 7. Translation results of cross-attention cGAN (XAcGAN) model with attention module location “0011.” Column (a): middle slices of input bright-field image stacks; column (b): ground truth fluorescent images, with nuclei false-colored such that magenta represents healthy nuclei and green represents apoptotic nuclei; column (c): translation results from the model with equivalent false-coloring applied; column (d): the ground truth classification of nuclei, gray represents healthy nuclei and white represents apoptotic nuclei; column (e): the semantic segmentation results by XAcGAN 0011 model; column (f): the MAE error maps between the target and generative fluorescent images.

5.1. Contribution of attention-based module

Compared to the pixel-to-pixel model, the attention-based model had higher accuracy in detecting nuclei. This proves that the attention element contributed to the long-distance dependency analysis when detecting nuclei from cell images (see Figure 8). Cells took up larger areas on bright-field images than nuclei in fluorescent images and included latent information on the nuclei’s state of health. However, without an attention-based mechanism, this information was restricted because of the sizes of convolutional kernels, which failed to transfer them to higher-level feature maps. To promote accuracy in nuclei detection, the attention-based module from the XAcGAN model played a vital role.

Figure 8. Detection accuracy comparison. (a) Ground-truth fluorescent image, (b) XAcGAN model result, (c) pixel-to-pixel model result. Translation result from the XAcGAN model has a higher accuracy of nuclei detection than the non-attention model.

5.2. XAcGAN model for necrotic cell detection

The cross-attention model has advantages in finding necrotic cells which are inconspicuous in bright-field images and out-of-focus in fluorescent images. Apoptosis is described as an energy-dependent process which is coordinated by cysteine-aspartic proteases called caspases, while necrosis is referred to as uncontrolled cell death(Reference Festjens, vanden Berghe and Vandenabeele 57 ). Necrosis is an energy-independent process and is a consequence of severe and sudden cellular damage to the extent that the cell is no longer functional. A notable difference in necrosis is the morphology of the nuclei. Within necrotic cells the nuclei undergo karyolysis when the nuclear material dissolves in the cytoplasm (broken down by endonucleases)(Reference Elmore 7 , Reference D’Arcy 58 ), which is observed in Figure 9. In Figure 9, we can see that the reference model failed to recognize the necrotic cells, but the cross-attention model found it, though the area of the target predicted by the cross-attention cGAN model was smaller than expected. This is an illustration of the contribution of the mask generation path to image translation. On the input image stacks, the necrotic cells were easily overlooked as the intensity changes in this area were not distinct. The mask generation path had better performance in indicating the location of these blurry items and leading the image generation path to predicting intensity.

Figure 9. Translation result of the necrotic cell. (a) Bright-field image. (b) Ground-truth fluorescent image. (c) Ground-truth nuclei classification result. (d) Result from the model without cross-attention module. (e) Result from XAcGAN model. (f) Nuclei segmentation result from XAcGAN model. Yellow circles indicate the necrotic cell under karyolysis.

5.3. Influence of the number of slices of bright-field image stack

We verified the performance of the XAcGAN model with a different number of input bright-field image slices. In the work of Christiansen et al. (Reference Christiansen, Yang and Ando 11 ), they found if the number of input slices was higher than five, the performance of the model did not improve too much. We performed a similar test on our model, using 1, 3, 5, and 7 slices. An odd number of slices was used for all tests since this maintained the same middle layer for each image stack. The input bright-field stack had 13 slices, the space distance of adjacent slices was 0.3 μm and the total depth of the stack was 3.6 μm. For input images with 3, 5, and 7 slices, the test was applied using two approaches. In the first, inter-slice separations were kept at 0.3 μm, thus yielding stack depths of 0.6 μm, 1.2 μm, and 1.8 μm, respectively. In the second test, slice separations of 1.8 μm, 0.9 μm, and 0.6 μm were used to produce fixed stack depths of 3.6 μm. The results of the tests are shown in Table 2.

Table 2. Performance of cross-attention cGAN (XAcGAN) model.

Abbreviations: PSNR, peak signal-to-noise ratio; SSIM, structural similarity index measure.

From Table 2, we can see that the input image with only one slice had the worst performance. Undoubtedly, the more slices the bright-field image stack had, the better result obtained. Meanwhile, the depth of the bright-field image stack had more influence on the translation result. For instance, from Figure 10, for a stack with three slices, performance for stack depth 3.6 μm and slice-separation 1.8 μm was significantly better than for stack depth 0.6 μm and slice-separation 0.3 μm. The reason for this could be the closer slices contained less 3D space information and led to poorer performance. As the number of slices of the input image stack increased, the improvement of the translation result was limited.

Figure 10. Performance of different numbers of input slices of bright-field image stacks, the numbers at the bottom indicate the number of image slices, “s” means slice separation remains unchanged, and “d” represents total depth unchanged.

6. Ablation Study

The cross-attention module was inspired by the self-attention module from the SAGAN model(Reference Zhang, Goodfellow, Metaxas and Odena 27 ). Therefore, we compared the performance of these two modules. As the self-attention module took one stack of feature maps as input, the self-attention cGAN (SAcGAN) model for fluorescent image translation contained only one independent generation path. We first applied the SAcGAN model only for translating fluorescent images without outputting the nuclei state evaluation. In this case, the model had one output channel and nuclei with different states were presented in one image. We then tried to train the SAcGAN model with two output channels aimed to test the performance of nuclei classification without the help of a mask generation path.

For the SAcGAN model with one output channel, the architecture of the SAcGAN model could be seen as the XAcGAN model without the mask generation path and with the cross-attention modules replaced with self-attention modules. The structure for each sub-module and strategies for choosing the inserting position were the same. The performance of the SAcGAN model is listed in Table 3. From Table 3, we found the SAcGAN models did significantly better than the pixel-to-pixel model; among SAcGAN models, model 0011 provided the best performance in all three evaluation metrics. Figure 11 shows the fluorescent image translation result of the SAcGAN model 0011.

Table 3. Performance of self-attention cGAN (SAcGAN) model.

Abbreviations: PSNR, peak signal-to-noise ratio; SSIM, structural similarity index measure.

Figure 11. Translation results and comparison of self-attention cGAN (SAcGAN) model. (a) Middle slices of input bright-filed image stacks. (b) Ground-truth fluorescent images. (c) Results from pixel-to-pixel model. (d) Results from SAcGAN model (0011). (e) Results from XAcGAN model (0011).

Statistically speaking, compared to the SAcGAN model, the result of the XAcGAN model which contained mask generation was not overwhelmingly better through the evaluation metrics. However, the function of predicting the state of cells or nuclei was irreplaceable and this is the main advantage of our model. To evaluate the accuracy of nuclear detection and classification, the state of nuclei in the ground-truth fluorescent images was recorded and compared to detections in the translated images generated by the tested models. For all models, possible translation errors covered unexpected nuclei (false positive) and undetected nuclei (false negative). While for XAcGAN, it was also possible to evaluate the misclassification of nuclei (“healthy” or “apoptotic”), for the single output channel models (Pix2Pix and SAcGAN), this had to be done manually based on the evaluation of nuclear texture. In practice, nuclei from single output channel models were frequently difficult to classify in this manner due to ambiguous textures. As such, the reliability of counts for Pix2Pix and SAcGAN might not be highly accurate in Table 4 as it is affected by human factors but reflects the consequence of the no-mask-assistance model.

Table 4. Nuclei classification result.

a “H to A” represents healthy nuclei predicted to be apoptotic.

b “A to H” represents apoptotic nuclei predicted to be healthy.

Table 4 contains the detection and classification result of the XAcGAN 0011 model: the nuclei state prediction accuracy was 85.68%; the false positive error was 1.27%, the false negative error was 4.78%, and the state classification error for detected nuclei was 8.27%, which is comprised of a healthy-to-apoptotic error of 4.01% and apoptotic-to-healthy error of 4.26%. From Table 4, it is evident that the XAcGAN model performed significantly better. One reason was that our model was trained on the database with a reliable classification ground truth. This advantage came from the process of dataset pre-processing, in which we used the whole 13 slices of the fluorescent image stacks for mask generation. For the models without a predicted mask from the raw inputs, to perform state prediction, additional methods for classification were required. However, the classification was applied to predicted outputs, so it would be hard to guarantee the accuracy of the classification.

We also tested the performance of SAcGAN with multiple output channels. In this test, each channel was aimed to present one state of nuclei, healthy or apoptotic. From Figure 12, we found that the model failed to produce fluorescent images. The results showed that the SAcGAN model tended to translate nuclei into one channel and ignore the generation in the other channel. For example, in Figure 12, only potential apoptotic nuclei were translated, and all the healthy nuclei were ignored. On the contrary, other models only recognized healthy nuclei. The reason the SAcGAN model was unable to process the classification could be that the loss function of the image generation task was too weak to lead the model to promote the classification result. To reduce the losses, the model tended to sacrifice one channel’s output and get stuck in a local minimum. The mask generation path assisted the model in skipping the local minimum and guided the image translation path on which channel to translate the nuclei. This task was used as a comparison of the multi-output-channel model with and without a mask generation path, and to explore the necessity of information connection between the mask generation and image generation.

Figure 12. Translation results of self-attention cGAN model with two output channels (without mask generation path). (a,b) Middle slices of bright-field image stacks and corresponding ground-truth fluorescent images. (c–f) Results from the SAcGAN model which has no mask prediction path. (c) Model 0000 (no attention module applied). (d) Model 0001. (e) Model 0011. (f) Model 0111. Magenta nuclei indicate healthy nuclei and green nuclei indicate apoptotic nuclei.

7. Conclusion

Our XAcGAN model achieved excellent performance in bright-field to fluorescent image translation tasks and provided promising health state prediction simultaneously. cGANs are a powerful tool in supervised image translation. The shortage of datasets didn’t affect the performance of our network. Attention-based modules applied in our model improved the performance of the image translation. Our model does not need multiple spatial extents as input, the attention modules encouraged the model to learn the long-distance spatial dependencies. Meanwhile, the cross-attention modules combined dependencies from both generation paths. The segmentation path was a crucial auxiliary for multi-state fluorescent image translation, as it guided the network on which channels the items to be generated should be assigned to. In addition, the prediction result can be used as a supplementary for healthy state evaluation.

For the biological research of monitoring the health state of cells and nuclei over time, the model is an alternative to the time and labor-consuming process of fluorescent labeling. It will reduce the work for microscopy experiments where only bright-field imaging is adequate for nuclei observation. This would be the most significant contribution of our research to cellular biology studies.

Acknowledgments

We would like to thank all the people from Bristol VI-Lab for their positive input and fruitful discussions during this project. We thank the technical staff from the Wolfson Bioimaging Facility for their expert support.

Authorship contribution

Conceptualization: R.W., P.V., A.A., S.C.; Data acquisition: D.B., S.C.; Methodology: R.W.; Model coding and training: R.W.; Writing original draft: R.W., D.B.; Writing revisions: P.V., A.A., S.C. All authors approved the final submitted draft.

Competing interest

The authors declare no competing interests exist.

Funding statement

D.B. was supported by a grant from the EPSRC CDT (EP/L015218/1). A.A. was supported in part by the U.S. National Institute of Health (NIH) via grant 1R01GM143388-01A1. The confocal microscope used was funded by BBSRC Alert 13 capital grant (BB/L014181/1).

Data availability statement

Code for training and for using the pre-trained model for translation is available on GitHub at https://github.com/SpikeRXWong/fluorescent_image_translation.git. All the images used in this work can be accessed at https://doi.org/10.5523/bris.2w8trsx55b0k22qez7uycg9sf4.

References

Croft, WJ (2006) Under the Microscope: A Brief History of Microscopy. Singapore: World Scientific.CrossRefGoogle Scholar
Lakowicz, JR (2006) Principles of Fluorescence Spectroscopy. Boston, MA: Springer.CrossRefGoogle Scholar
Murphy, DB (2002) Fundamentals of Light Microscopy and Electronic Imaging, 2nd ed. Boston, MA: John Wiley & Sons.Google Scholar
Jia, HR, Zhu, YX & Wu, FG (2020) Introduction: fluorescent materials for cell imaging. In Fluorescent Materials for Cell Imaging, pp. 115 [Wu, FG, editor]. Singapore: Springer.Google Scholar
Andersson, H, Baechi, T, Hoechl, M & Richter, C (1998) Autofluorescence of living cells. J Microscopy 191(Pt 1), 17.CrossRefGoogle ScholarPubMed
Bucevičius, J, Lukinavičius, G & Gerasimaitė, R (2018) The use of Hoechst dyes for DNA staining and beyond. Chemosensors 6(2), 18.CrossRefGoogle Scholar
Elmore, S (2007) Apoptosis: a review of programmed cell death. Toxicol Pathol 35(4), 495516.CrossRefGoogle ScholarPubMed
Nikolai, V (2011) Binding of Hoechst with nucleic acids using fluorescence spectroscopy. J Biophys Chem 2, 443447.Google Scholar
Laissue, PP, Alghamdi, RA, Tomancak, P, Reynaud, EG & Shroff, H (2017) Assessing phototoxicity in live fluorescence imaging. Nat Methods 14(7), 657661.CrossRefGoogle ScholarPubMed
Waters, JC (2013) Live-cell fluorescence imaging. In Methods in cell biology (Sluder, G & Wolf, DE, Eds.), 114, 125150. Cambridge, Massachusetts: Academic Press.Google Scholar
Christiansen, EM, Yang, SJ, Ando, DM, et al. (2018) In silico labeling: predicting fluorescent labels in unlabeled images. Cell 173(3), 792803.e19.CrossRefGoogle ScholarPubMed
LeCun, Y, Bengio, Y & Hinton, G (2015) Deep learning. Nature 521(7553), 436444.CrossRefGoogle ScholarPubMed
Goodfellow, I, Bengio, Y & Courville, A (2016) Deep learning. Cambridge, Massachusetts: MIT press.Google Scholar
Rawat, W & Wang, Z (2017) Deep convolutional neural networks for image classification: a comprehensive review. Neural Comput 29(9), 23522449.CrossRefGoogle ScholarPubMed
Moen, E, Bannon, D, Kudo, T, Graf, W, Covert, M & van Valen, D (2019) Deep learning for cellular image analysis. Nat Methods 16(12), 12331246.CrossRefGoogle ScholarPubMed
Goodfellow, I, Pouget-Abadie, J, Mirza, M, et al. (2020) Generative adversarial networks. Commun ACM 63 (11), 139144.CrossRefGoogle Scholar
Karras, T, Aila, T, Laine, S & Lehtinen, J (2018) Progressive growing of GANs for improved quality, stability, and variation. In International Conference on Learning Representations. Vancouver, Canada. Retrieved from https://openreview.net/forum?id=Hk99zCeAb¬eId=Hk99zCeAb&ref=https://githubhelp.com.Google Scholar
Zhu, JY, Park, T, Isola, P & Efros, AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, pp. 22232232. Cambridge, MA: IEEE Press.Google Scholar
Taigman, Y, Polyak, A & Wolf, L (2017) Unsupervised cross-domain image generation. In International Conference on Learning Representations. Toulon, France. Retrieved from https://openreview.net/forum?id=Sk2Im59ex.Google Scholar
Isola, P, Zhu, JY, Zhou, T & Efros, AA (2017) Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 11251134. Cambridge, MA: IEEE Press.Google Scholar
Li, X, Zhang, G, Qiao, H, et al. (2021) Unsupervised content-preserving transformation for optical microscopy. Light: Sci Appl 10(1), 111.CrossRefGoogle ScholarPubMed
Ounkomol, C, Seshamani, S, Maleckar, MM, Collman, F & Johnson, GR (2018) Label-free prediction of three-dimensional fluorescence images from transmitted-light microscopy. Nat Methods 15(11), 917920.CrossRefGoogle ScholarPubMed
Nguyen, TC, Bui, V, Thai, A, et al. (2020) Virtual organelle self-coding for fluorescence imaging via adversarial learning. J Biomed Optics 25(9), 096009.CrossRefGoogle ScholarPubMed
Lee, G, Oh, JW, Kang, MS, Her, NG, Kim, MH & Jeong, WK (2018) DeepHCS: bright-field to fluorescence microscopy image conversion using deep learning for label-free high-content screening. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16–20, 2018, Proceedings, vol. 2(11), pp. 335343. New York: Springer International Publishing.CrossRefGoogle Scholar
Zhang, G, Ning, B, Hui, H, et al. (2022) Image-to-images translation for multiple virtual histological staining of unlabeled human carotid atherosclerotic tissue. Mol Imaging Biol 24(1), 3141.CrossRefGoogle ScholarPubMed
Vaswani, A, Shazeer, N, Parmar, N, et al. (2017) Attention is all you need. In Advances in Neural Information Processing Systems, vol. 30 [Guyon, I, Von Luxburg, U, Bengio, S, et al., editors]. Cambridge, MA: MIT Press.Google Scholar
Zhang, H, Goodfellow, I, Metaxas, D & Odena, A (2019) Self-attention generative adversarial networks. In International Conference on Machine Learning. Proceedings of Machine Learning Research, pp. 73547363. Cambridge, MA: Journal of Machine Learning Research.Google Scholar
Brent, R & Boucheron, L (2018) Deep learning to predict microscope images. Nat Methods, 15(11), 868870.CrossRefGoogle ScholarPubMed
Ronneberger, O, Fischer, P & Brox, T (2015) U-Net: convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234241. Cham: Springer.Google Scholar
Szegedy, C, Liu, W, Jia, Y, et al. (2015) Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 19. Cambridge, MA: IEEE Press.Google Scholar
Denton, EL, Chintala, S & Fergus, R (2015) Deep generative image models using a Laplacian pyramid of adversarial networks. In Advances in Neural Information Processing Systems, vol. 28 [Cortes, C, Lawrence, N, et al., editors]. Cambridge, MA: MIT Press.Google Scholar
Radford, A, Metz, L & Chintala, S (2015) Unsupervised representation learning with deep convolutional generative adversarial networks. Preprint, arXiv:1511.06434.Google Scholar
Salimans, T, Goodfellow, I, Zaremba, W, Cheung, V, Radford, A & Chen, X (2016) Improved techniques for training GANs. In Advances in Neural Information Processing Systems, vol. 29 [Lee, D, Sugiyama, M, et al., editors]. Cambridge, MA: MIT Press.Google Scholar
Gui, J, Sun, Z, Wen, Y, Tao, D & Ye, J (2023) A review on generative adversarial networks: algorithms, theory, and applications. In IEEE Transactions on Knowledge and Data Engineering, vol. 35(4), pp. 33133332. Cambridge, MA: IEEE Press.Google Scholar
Yi, X, Walia, E & Babyn, P (2019) Generative adversarial network in medical imaging: a review. Med Image Anal 58, 101552.CrossRefGoogle ScholarPubMed
Mirza, M & Osindero, S (2014) Conditional generative adversarial nets. Preprint, arXiv:1411.1784.Google Scholar
Li, C & Wand, M (2016) Precomputed real-time texture synthesis with Markovian generative adversarial networks. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, vol. 3(14), pp. 702716. New York: Springer International Publishing.CrossRefGoogle Scholar
Hollandi, R, Szkalisity, A, Toth, T, et al. (2020) nucleAIzer: a parameter-free deep learning framework for nucleus segmentation using image style transfer. Cell Syst 10(5), 453458.CrossRefGoogle ScholarPubMed
Han, L & Yin, Z (2017) Transferring microscopy image modalities with conditional generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 99107. Cambridge, MA: IEEE Press.Google Scholar
Lee, G, Oh, JW, Her, NG & Jeong, WK (2021) Deephcs++: bright-field to fluorescence microscopy image conversion using multi-task learning with adversarial losses for label-free high-content screening. Med Image Anal 70, 101995.CrossRefGoogle ScholarPubMed
Cordonnier, JB, et al. (2020) On the relationship between self-attention and convolutional layers. In International Conference on Learning Representations. Addis Ababa, Ethiopia. Retrieved from https://openreview.net/forum?id=HJlnC1rKPB.Google Scholar
Cheng, J, Dong, L & Lapata, M (2016) Long short-term memory networks for machine reading. In EMNLP 2016: Conference on Empirical Methods in Natural Language Processing, pp. 551562. Cambridge, MA: MIT Press.CrossRefGoogle Scholar
Parmar, N, Vaswani, A, Uszkoreit, J, et al. (2018) Image transformer. In International Conference on Machine Learning. Proceedings of Machine Learning Research, pp. 40554064. Cambridge, MA: Journal of Machine Learning Research.Google Scholar
Guo, MH, Xu, TX, Liu, JJ, et al. (2022) Attention mechanisms in computer vision: a survey. Comput Visual Media 8, 331368.CrossRefGoogle Scholar
Miyato, T, Kataoka, T, Koyama, M & Yoshida, Y (2018) Spectral normalization for generative adversarial networks. In International Conference on Learning Representations. Vancouver, Canada. Retrieved from https://openreview.net/forum?id=B1QRgziT-.Google Scholar
Miyato, T & Koyama, M (2018) cGANs with projection discriminator. Preprint, arXiv:1802.05637.Google Scholar
Odena, A, Buckman, J, Olsson, C, et al. (2018) Is generator conditioning causally related to GAN performance? In International Conference on Machine Learning. Proceedings of Machine Learning Research, pp. 38493858. Cambridge, MA: Journal of Machine Learning Research.Google Scholar
He, K, Zhang, X, Ren, S & Sun, J (2016) Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770778. Cambridge, MA: IEEE Press.Google Scholar
Hou, R, et al. (2019) Cross attention network for few-shot classification. In Advances in Neural Information Processing Systems, 32.Google Scholar
Hou, R, Chang, H, Ma, B, Shan, S, & Chen, X (2019) Cross attention network for few-shot classification. In Advances in Neural Information Processing Systems, vol. 32 [Wallach, H, Larochelle, H, et al., editors]. Cambridge, MA: MIT Press.Google Scholar
Pizer, SM, Johnston, RE, Ericksen, JP, Yankaskas, BC, Muller, KE & Medical Image Display Research Group (1990) Contrast-limited adaptive histogram equalization: speed and effectiveness. In Proceedings of the First Conference on Visualization in Biomedical Computing, Atlanta, Georgia, vol. 337, p. 1. Cambridge, MA: IEEE Press.Google Scholar
Beucher, S & Lantuejoul, C (1979). Use of watersheds in contour detection. In Proc. Int. Workshop on Image Processing, Sept. 1979, 1721Google Scholar
Beucher, S (1982) Watersheds of functions and picture segmentation. In ICASSP’82. IEEE International Conference on Acoustics, Speech, and Signal Processing, pp. 19281931. Cambridge, MA: IEEE Press.CrossRefGoogle Scholar
Otsu, N (1979) A threshold selection method from gray-level histograms. IEEE Trans Syst Man Cybernetics 9(1), 6266.CrossRefGoogle Scholar
Kingma, DP & Ba, J (2014) Adam: a method for stochastic optimization. Preprint, arXiv:1412.6980.Google Scholar
Heusel, M, Ramsauer, H, Unterthiner, T, Nessler, B & Hochreiter, S (2017) GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems, vol. 30 [Guyon, I, Von LuxBurg, U, et al., editors]. Cambridge, MA: MIT Press.Google Scholar
Festjens, N, vanden Berghe, T & Vandenabeele, P (2006) Necrosis, a well-orchestrated form of cell demise: signalling cascades, important mediators and concomitant immune response. Biochimica et Biophysica Acta (BBA)-Bioenergetics 1757(9–10), 13711387.CrossRefGoogle ScholarPubMed
D’Arcy, MS (2019) Cell death: a review of the major forms of apoptosis, necrosis and autophagy. Cell Biol Int 43(6), 582592.CrossRefGoogle ScholarPubMed
Figure 0

Figure 1. Information comparison between bright-field light microscopy (left) with fluorescence light microscopy (right) of the same cells. Green boxes indicate zoomed areas of healthy nuclei and magenta boxes indicate apoptotic nuclei with yellow arrows pointing to the position of the cells in the overview images.

Figure 1

Figure 2. Network architectures of sub-modules for generator. (a) Down-sampling sub-module. (b) Bottleneck sub-module. (c) Up-sampling sub-module.

Figure 2

Figure 3. Generator architecture. The numbers in each sub-module indicate the numbers of input and output channels, respectively. Architectures of sub-modules are presented in Figure 2, and attention modules are shown in Figure 4.

Figure 3

Figure 4. Attention module architectures. (a) Architecture of the self-attention module. (b) Architecture of the cross-attention module.

Figure 4

Figure 5. Discriminator architecture. The numbers in each sub-module indicate the numbers of input and output channels, respectively. Numbers in attention boxes represent the size of input feature maps.

Figure 5

Figure 6. Dataset preparation process. (a) Maximum intensity along z-stack of fluorescent images, (b) threshold and watershed segmentation output, (c) automatic classification result, (d) manually revised result, the revised individual in the yellow circle, (e) example of image dataset for training, contains bright-field images, split fluorescent images, and masks.

Figure 6

Table 1. Performance of cross-attention conditional GAN model.

Figure 7

Figure 7. Translation results of cross-attention cGAN (XAcGAN) model with attention module location “0011.” Column (a): middle slices of input bright-field image stacks; column (b): ground truth fluorescent images, with nuclei false-colored such that magenta represents healthy nuclei and green represents apoptotic nuclei; column (c): translation results from the model with equivalent false-coloring applied; column (d): the ground truth classification of nuclei, gray represents healthy nuclei and white represents apoptotic nuclei; column (e): the semantic segmentation results by XAcGAN 0011 model; column (f): the MAE error maps between the target and generative fluorescent images.

Figure 8

Figure 8. Detection accuracy comparison. (a) Ground-truth fluorescent image, (b) XAcGAN model result, (c) pixel-to-pixel model result. Translation result from the XAcGAN model has a higher accuracy of nuclei detection than the non-attention model.

Figure 9

Figure 9. Translation result of the necrotic cell. (a) Bright-field image. (b) Ground-truth fluorescent image. (c) Ground-truth nuclei classification result. (d) Result from the model without cross-attention module. (e) Result from XAcGAN model. (f) Nuclei segmentation result from XAcGAN model. Yellow circles indicate the necrotic cell under karyolysis.

Figure 10

Table 2. Performance of cross-attention cGAN (XAcGAN) model.

Figure 11

Figure 10. Performance of different numbers of input slices of bright-field image stacks, the numbers at the bottom indicate the number of image slices, “s” means slice separation remains unchanged, and “d” represents total depth unchanged.

Figure 12

Table 3. Performance of self-attention cGAN (SAcGAN) model.

Figure 13

Figure 11. Translation results and comparison of self-attention cGAN (SAcGAN) model. (a) Middle slices of input bright-filed image stacks. (b) Ground-truth fluorescent images. (c) Results from pixel-to-pixel model. (d) Results from SAcGAN model (0011). (e) Results from XAcGAN model (0011).

Figure 14

Table 4. Nuclei classification result.

Figure 15

Figure 12. Translation results of self-attention cGAN model with two output channels (without mask generation path). (a,b) Middle slices of bright-field image stacks and corresponding ground-truth fluorescent images. (c–f) Results from the SAcGAN model which has no mask prediction path. (c) Model 0000 (no attention module applied). (d) Model 0001. (e) Model 0011. (f) Model 0111. Magenta nuclei indicate healthy nuclei and green nuclei indicate apoptotic nuclei.