Chinese Optics Letters, Volume. 20, Issue 3, 031701(2022)

Fluo-Fluo translation based on deep learning

Zhengfen Jiang1, Boyi Li2, Tho N. H. T. Tran2, Jiehui Jiang1, Xin Liu2,3、*, and Dean Ta2,4、**
Author Affiliations
  • 1School of Communication & Information Engineering, Shanghai University, Shanghai 200444, China
  • 2Academy for Engineering & Technology, Fudan University, Shanghai 200433, China
  • 3State Key Laboratory of Medical Neurobiology, Fudan University, Shanghai 200433, China
  • 4Center for Biomedical Engineering, Fudan University, Shanghai 200433, China
  • show less
    Figures & Tables(8)
    System for Fluo-Fluo translation based on cGAN. (A) The training dataset is composed of fluorescence images x and y in the same field of view. (B) A deep neural network is composed of untrained parameters. (C) The deep neural network trained with data in (A). (D) Test image. (E) Based on the trained deep learning model, the fluorescence image y is predicted from the fluorescence image x.
    cGAN framework for Fluo-Fluo translation based on deep learning. (A) The generator network attempts to generate image y with respect to image x, and the discriminator network attempts to distinguish between the generated image y and the true image y. There is a competitive relationship between these two networks. Briefly, image x is used as the input of the generator to obtain the generated image G(x), and then G(x) and x are combined as the input of the discriminator. During training, two error functions are calculated: (i) L1 and MS-SSIM loss functions are used to measure the similarity between the generated image G(x) and the target image y; (ii) the cGAN error attempts to distinguish the generated image G(x) from the target image y corresponding to the input image x. The combined loss functions are optimized by the Adam algorithm. (B) Once trained, the generator can immediately predict the fluorescence image y from the fluorescence image x of the test dataset.
    Prediction results obtained by the proposed method. (A) Predict TuJ1 protein from Islet1 protein based on deep learning. (B) Predict CellMask from DAPI based on deep learning. (C) Predict PI from Hoechst based on deep learning. From left to right, the input image, the true image (ground truth), the network generated image, the absolute error map, and the scatter plot are displayed in turn.
    Prediction results obtained by the proposed method. (A) Predict MAP2 from DAPI based on deep learning. (B) Predict NFH from DAPI based on deep learning. (C) Multi-label visualization co-localization.
    Effects of the number of training sets on the prediction performance of the proposed method. (A) and (B) Prediction results obtained by using the reduced training sets. (A) Predict TuJ1 from Islet1. (B) Predict PI from Hoechst. (C) and (D) Quantify the differences in terms of the SSIM, PSNR, and MAE indicators, respectively.
    Predict CellMask from DAPI based on our rat cardiomyocyte dataset. The public dataset from Group 2 (predicting CellMask from DAPI based on human breast cancer cells) is used for network training, and then the trained model is used to predict our own experimental data.
    • Table 1. Detailed Information of Experimental Data

      View table
      View in Article

      Table 1. Detailed Information of Experimental Data

      GroupsCell TypeFluorescence LabelMarked LocationTraining SetTest Set
      Group 1Human motor neuronsIslet1Motor neurons1280320
      TuJ1Neurons
      Group 2Human breast cancer cellsDAPINuclei30075
      CellMaskMembrane
      Group 3Rat cortical neuronsHoechstNuclei57601440
      PIDead cells
      Group 4Human motor neuronsDAPINuclei2400600
      MAP2Dendrites
      NFHAxons
    • Table 2. SSIM, PSNR, and MAE Values Between the True and the Network Generated Images

      View table
      View in Article

      Table 2. SSIM, PSNR, and MAE Values Between the True and the Network Generated Images

      ComparisonsSSIMPSNR (dB)MAE
      MeanStdMeanStdMeanStd
      TUJ1 (true) vs TUJ1 (generated)0.8020.02421.8450.8215.6821.468
      CellMask (true) vs CellMask (generated)0.8490.02823.7320.9486.3481.232
      PI (true) vs PI (generated)0.9800.00929.4563.2660.8850.610
      MAP2 (true) vs MAP2 (generated)0.8880.03023.1721.5953.9991.821
      NFH (true) vs NFH (generated)0.7290.06518.1301.5364.6161.638
    Tools

    Get Citation

    Copy Citation Text

    Zhengfen Jiang, Boyi Li, Tho N. H. T. Tran, Jiehui Jiang, Xin Liu, Dean Ta. Fluo-Fluo translation based on deep learning[J]. Chinese Optics Letters, 2022, 20(3): 031701

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Biophotonics

    Received: Nov. 11, 2021

    Accepted: Dec. 13, 2021

    Posted: Dec. 14, 2021

    Published Online: Jan. 11, 2022

    The Author Email: Xin Liu (xinliu.c@gmail.com), Dean Ta (tda@fudan.edu.cn)

    DOI:10.3788/COL202220.031701

    Topics