Journal of Healthcare Engineering

Journal of Healthcare Engineering / 2021 / Article
Special Issue

Multiple Criteria Decision-Making Approaches for Healthcare Management Applications

View this Special Issue

Research Article | Open Access

Volume 2021 |Article ID 9917545 | https://doi.org/10.1155/2021/9917545

Ching-Chun Chang, Xu Wang, Ji-Hwei Horng, Isao Echizen, "Progressive Transmission of Medical Images via a Bank of Generative Adversarial Networks", Journal of Healthcare Engineering, vol. 2021, Article ID 9917545, 13 pages, 2021. https://doi.org/10.1155/2021/9917545

Progressive Transmission of Medical Images via a Bank of Generative Adversarial Networks

Academic Editor: Chi-Hua Chen
Received13 Mar 2021
Revised29 Mar 2021
Accepted09 Apr 2021
Published28 Apr 2021

Abstract

The healthcare sector is currently undergoing a major transformation due to the recent advances in deep learning and artificial intelligence. Despite a significant breakthrough in medical imaging and diagnosis, there are still many open issues and undeveloped applications in the healthcare domain. In particular, transmission of a large volume of medical images proves to be a challenging and time-consuming problem, and yet no prior studies have investigated the use of deep neural networks towards this task. The purpose of this paper is to introduce and develop a deep-learning approach for the efficient transmission of medical images, with a particular interest in the progressive coding of bit-planes. We establish a connection between bit-plane synthesis and image-to-image translation and propose a two-step pipeline for progressive image transmission. First, a bank of generative adversarial networks is trained for predicting bit-planes in a top-down manner, and then prediction residuals are encoded with a tailored adaptive lossless compression algorithm. Experimental results validate the effectiveness of the network bank for generating an accurate low-order bit-plane from high-order bit-planes and demonstrate an advantage of the tailored compression algorithm over conventional arithmetic coding for this special type of prediction residuals in terms of compression ratio.

1. Introduction

With the development of digital imaging equipment, more and more medical images have been produced and a medical image dataset usually contains a large number of images. Transmission of abundant medical images over a low narrow bandwidth public network will no doubt cause critical network pressure and further delay the timing of diagnosis [1].

To increase the speed of medical image transmission, some schemes have been proposed to build up more effective special networks. For instance, in [2], a unique pair of interfaces was proposed to provide a fast thoroughfare for medical data transmission between two DICOM applications and could speed up 1.22 and 13 times, respectively, in the local area networks (LAN) and wide area networks (WAN). Later, they further proposed an improved network [3] to combine the local area network (LAN) and wide area network (WAN) and used parallel TCP connections to optimize the DICOM protocol. This network can speed up about 2.2 to 3.5 times when transmitting magnetic resonance images.

On the other aspect, some new image transmission techniques have been proposed. An intuitive way is to reduce the resolution of medical images before transmission in a poor bandwidth network. However, doctors suffer from misdiagnosis with the lower resolution medical images. Therefore, the progressive image transmission (PIT) [4] technique emerges as the times require. The conventional raster scan ordered pixel by pixel image transmission technique is inefficient, and the key region of the image which indicates the disease may be received at the end of the transmission makes doctors out of patience and affects the treatment. The progressive image transmission technique usually transmits a lower resolution image at first and gradually transmits the details of the image to further increase the resolution until the original image is totally transmitted. Through the transmission, receivers can decide whether they want to receive the details of the image or not in a short time. As a consequence, doctors can control the final image resolution and the required storage space. Over the past decades, some mature PIT techniques have been proposed. The bit-plane progressive image transmission method (BPM) [5, 6] is one of the low computational complexity and intuitive technique to transmit the image bit-plane by bit-plane which is detailed in Section 2. Another sort of progressive image transmission techniques is that, in [712], the image is divided into nonoverlapping blocks, and a quantized pixel value is used to represent the whole block, then, progressively dividing the image into finer blocks to improve the image resolution. Some schemes [1317] transform the image into a wavelet domain and transmit the coefficients instead of pixels. Lossy compression [18] and lossless compression [19] methods are also adopted in some schemes to reduce the file size in each round of transmission.

The deep learning technique has matured in recent years and drawn massive research which is widely used in industry [20]. Some advanced deep learning architectures have also been introduced into medical image processing. Medical image classification is the most important area where deep learning makes a great contribution. The CNN network has been widely used in medical image classification and achieved a higher accuracy [2123]. The deep learning technique is also suitable for medical image detection. The methods for organ, region, and landmark localization [2426] and object and lesion detection [2729] all adopt the deep learning architectures. Then, the medical image segmentation is naturally developed in the detected images [3032]. The other tasks on medical images processing such as content-based image retrieval [33]; image generation and enhancement [34]; and description of medical images with shape, margin, and density [35] have also been proposed. Recently, the deep learning architecture with adversarial learning: generative adversarial network (GAN) has been proposed by Goodfellow et al. [36] that has the ability to generate realistic-looking images. Based on the GAN network, a powerful architecture pix2pix [37] was set up as an image bit-plane predictor to translate the image styles and has been adopted in different scenarios such as steganography [38] and even in medical images [39, 40]. However, so far, these deep learning techniques have not been applied to the medical image transmission. By leveraging the prediction ability of the GAN network, the medical images can be compressed before transmission and thus the efficiency of transmission can be improved.

In this paper, we adopt the framework of the bit-plane PIT method to transmit medical images. To reduce the amount of data before transmission, a GAN-based predictor bank is designed to predict the bit-planes. By progressively predicting the bit-plane to be sent based on previously transmitted bit-planes, we need only to send the residual provided that the receiver can execute the same prediction. The proposed GAN-based predictor bank produces a prediction of high accuracy so that the residual can be efficiently compressed before transmission. Thus, the efficiency of bit-plane PIT is greatly improved, and the transmission cost can be reduced. The main contributions of our scheme are summarized as follows:(1)Introduce a deep neural network of pix2pix GAN network into the research of progressive medical image prediction and transmission(2)Design a GAN-based predictor bank to help bit-plane compression of medical images(3)Propose an adaptive residual bit-plane compression technique of high compression ratio

Experimental results validate the effectiveness and efficiency of the proposed prediction network and compression technique. The rest of this paper is organized as follows. Section 2 briefly introduces the bit-plane method of the progressive image transmission. Section 3 details the proposed adversarial learning for progressive bit-plane prediction scheme and the adaptive residual bit-plane compression technique. Section 4 provides the experimental results to demonstrate the accuracy of our prediction scheme and the efficiency of the proposed compression technique. In addition, the performance of the proposed residual bit-plane compression is compared with the conventional arithmetic coding. The conclusions are offered in Section 5.

2. Bit-Plane Method of PIT

In this section, the bit-plane method of progressive image transmission (BPM) technique which has been widely used to transmit a large volume of high-resolution medical images, is briefly introduced. The progressive image transmission (PIT) technique is an effective way to transmit images over narrow bandwidth channels. Among the PIT techniques, BPM is a simple and intuitive way of transmission, which divides an image into bit-planes and sends it plane by plane. For the grayscale images, each pixel is valued from 0 to 255, that is, 256 gray levels, and can be represented by 8 bits. The original image can be decomposed into eight bit-planes from the most significant bit-plane b8 to the least significant bit-plane b1, so that b8 records the main content and b1 records the subtle details, respectively, of the original image. Therefore, in the bit-plane PIT method, bit-planes from b8 to b1 are sequentially transmitted and the receiver can progressively recover the image until all bit-planes are sent or the current image quality is satisfied. The detailed BPM method is described below.

In the most significant bit-plane b8, the bit “0” and “1” indicate the corresponding pixels are ranging from “0 to 127” and “128 to 255,” respectively. Therefore, after receiving b8, the mean values of these two intervals, that is, 64 and 192, are used to coarsely represent them. When receiving the next bit-plane b7, the receiver updates the image by using finer quantization levels of (32, 96, 160, 224) to represent corresponding codes “00,” “01,” “10,” and “11.” This process proceeds until all of the eight bit-planes are sent. An example of BPM transmission is illustrated in Figure 1, where the progressive combination of bit-planes and their corresponding recovered images are presented. As the number of received bit-planes increases, the available quantization levels also increase and a finer recovered image can be obtained.

3. Progressive Bit-Plane Prediction and Compression

Based on BPM progressive image transmission, we propose a bit-plane prediction method and a compression technique to improve the transmission efficiency of medical images. In this section, the proposed image transmission scheme and a pix2pix model-based bit-plane prediction method are introduced first. Then, a new technique for compression of the prediction error bit-plane is presented.

3.1. Adversarial Learning for Progressive Bit-Plane Prediction

For the purpose of progressively predicting image bit-plane by bit-plane, the generative adversarial network is introduced in this paper to predict and synthesize bit-planes. GAN architecture is a classical neural network to generate synthetic images of a specific style through training. The pix2pix model is a famous model of a GAN-based network which can be exploited to make an image-to-image translation. In the pix2pix model, the generator G is a U-Net, which is designed for biomedical image segmentation, instead of the conventional encoder-decoder net. The U-Net consists of a contracting path to downsample the input image, and an expansive path to generate an output image of full resolution. The contracting path consists of four steps connected by 2x2 max pooling layer with stride 2; each step contains three repeated 3 × 3 convolutions with ReLU. The expansive path also consists of four steps connected by 2 × 2 upconvolution and halves of the feature maps are cropped from the contracting path. For the discriminator, the PatchGAN is adopted to discriminate that the overlapping patches in the generated image are real or fake. Different from the conventional GAN discriminator which only output one result, that is, real or fake, PatchGAN outputs a matrix and each element reflects the pixel differences of the current patch between the synthetic image and the original image in the current patch region. Additionally, to obtain a generated image with higher similarity to the input original image, the Manhattan distance ( norm) which is more suitable for high dimensional images is chosen to evaluate the difference between input and output images. Therefore, in the pix2pix model, assume and are the input and output images; the difference between input and output is evaluated by

Then, the final objective is obtained by mixing the loss with the traditional GAN loss bywhere is a hyperparameter to balance two losses to achieve better performances.

Through the above architecture, the pix2pix model can translate the input image into different styles such as color images, day-to-night, and image inpainting. Inspired by pix2pix image-to-image translation model, in this paper, we train the pix2pix model to predict bit-planes of medical images. The proposed scheme is based on the BPM of progressive image transmission. A medical image is divided into eight bit-planes and transmitted plane by plane. On the receiver side, bit-planes are also received plane by plane and the medical image can be progressively reconstructed. During transmission, we leverage the pix2pix model to generate a predicted version of the bit-plane to be sent based on the previous bit-planes. Benefiting from the high accuracy of prediction, the prediction error of bit-plane (residual bit-plane) can be compressed efficiently. Thus, a bit-plane can be transmitted by sending its residual instead. The predictors of different layers of bit-plane are trained separately, organized into a predictor bank, and shared with the receiver. When receiving the residual, the receiver can do the same prediction to obtain the predicted bit-plane, and then, the original bit-plane can be recovered according to the received residual bit-plane and the predicted bit-plane. The framework of our proposed progressive medical image transmission scheme is shown in Figure 2.

In detail, assuming the bit-planes from the MSB to LSB are b8 to b1, we feed all b8 bit-planes of training images into the model as the input images to generate output images and set their corresponding b7 bit-planes as the ground truth images. The standard format of input and output for the pix2pix model is an image of 8-bit depth. To fit this format, zeros are appended behind the provided bit-planes. When training the model for predicting b6, b8 and b7 are concatenated as the input. An illustration for this case is shown in Figure 3. The detailed prediction models for the remaining bit-planes are trained in a similar manner. Then, a series of pretrained predictors are organized into a predictor bank as shown in Figure 4, which contains seven predictors for predicting bit-planes of different layers. For instance, the b6 predictor can predict bit-plane b6 by using the bit-planes b8 and b7. Therefore, each bit-plane from b7 to b1 can be progressively predicted by its previous bit-planes. On the application phase, the residual bit-plane can be calculated by comparing the bit-plane to be sent and the synthetic one bywhere , , and represent the residual, original, and synthetic bit-planes, respectively. Assuming the image is sized , denote the coordinates and range from 1 to . In this paper, the seven residual bit-planes are denoted by symbols r7 to r1. On the receiver side, the predicted bit-plane can be obtained by the preshared predictor bank. After receiving the residual bit-plane, the real bit-plane can be recovered bywhere denotes the binary bit-flipped value of , that is, change “1” to “0” or change “0” to “1.”

3.2. Adaptive Residual Bit-Plane Compression

The size of the prediction residual and its original bit-plane is the same. The major difference is that the values in the residual bit-plane are mostly zero. So, it can be compressed more efficiently. In addition, the error bits of our GAN-based bit-plane prediction are highly concentrated in the vicinity of image edges or complex details. According to the features of the residual bit-planes, we further propose an adaptive compression technique to reduce the data size. The detailed procedures are as follows:Step 1: decompose a residual bit-plane sized into nonoverlapping blocks sized , where ranges from 1 to .Step 2: initialize a location map with the length of .Step 3: scan the residual blocks in the raster scan order. If values in the block are all “0” or all “1,” go to Step 4; otherwise, go to Step 5.Step 4: set the -th bit of the location map to “0,” and record the block value by one bit, that is, , where denotes the -th segment of the recovery sequence and is the common value of the block’s elements.Step 5: set to “1,” and record all bits of the current residual block into the recovery sequence in raster scan order.Step 6: return to Step 3 until all blocks have been scanned to obtain the location map and the recovery sequence .Step 7: compress the location map by arithmetic coding to acquire the final location map . Then, the final compressed residual bit-plane is obtained by concatenating and .

When receiving the compressed residual bit-plane, the initial residual bit-plane can be obtained by the following decompression procedures:Step 1: decompress the location map by arithmetic decoding to obtain the sized location map .Step 2: if is “0,” all bits in the current block are the same and go to Step 3; otherwise, go to Step 4.Step 3: extract one-bit from the recovery sequence and set all bits in to .Step 4: extract bits from and rearrange these bits in the raster scan order to recover .Step 5: return to Step 2 until all blocks have been scanned. Then, the initial residual bit-plane can be obtained by tiling the recovered residual blocks in the raster scan order.

A simple example is given in Figure 5. In this example, the block size is set to 2. Since the first, third, and fourth blocks are uniform valued blocks, the corresponding bits in the location map are set to 0. Meanwhile, just one sample bit is recorded into the recovery sequence for each of them. For the rest blocks, their corresponding bits in the location map are set to 1 and all bit values are recorded. When recovering these blocks, referring to the location map and the recovery sequence, all blocks can be totally recovered.

4. Experimental Results

In this section, we evaluate the performance of our proposed bit-plane prediction scheme with generative adversarial networks and compare our proposed residual bit-plane compression scheme with the conventional arithmetic coding compression technique. The dataset used for training and testing, the training details, and the evaluation metrics are first introduced, and then, the performance of the proposed scheme is evaluated.

4.1. Dataset and Training Details

The image samples for training and testing are the greyscale chest X-ray images collected from the National Institutes of Health under the U.S. Department of Health and Human Services. In our experiments, 10,000 images from the first compressed file are used for training, and the other 10 images from the fifth compressed file as shown in Figure 6 are used for performance evaluating and analyzing. Before training and testing, all the images are first downsampled into 256 × 256 sized images to fit our proposed network architecture.

When training the proposed model, each bit-plane prediction is trained over 100 epochs and the initial learning rate and the batch size are set to and 32, respectively. The learning rate is enforced halfway during training and the hyperparameter is set to . The model parameters are updated and optimized by the Adam function.

4.2. Evaluation Metrics

To evaluate the performance of the predicted bit-planes, two important metrics, error rate (ER) and the compression ratio (CR), are introduced in this paper to conduct some experiments. For a predicted bit-plane, referring to its original bit-plane, if the two bits in the same place are the same, the corresponding bit in the residual bit-plane is set to “0” and set to “1” when the predicted bit is incorrect. The error rate (ER) is defined as the ratio of “1’s” in the residual bit-plane and can be calculated bywhere denotes the number of error bits. After obtaining the residual bit-plane, different compression techniques are adopted to compress it. The compression ratio (CR) denotes the ratio between the data sizes of the original and the compressed versions that can be calculated bywhere represents the data size, in bits, of the compressed residual. Generally, the compression ratio is greatly dependent on the error rate, and a residual of a low error rate can be compressed more efficiently. Thus, a prediction of high accuracy can improve the transmission efficiency.

4.3. Performances

First, the test image “Image 05” is used to demonstrate the experimental results of the original bit-planes, prediction bit-planes, and corresponding residual bit-planes.

From Figure 7, when receiving the bit-plane 8, the rest bit-planes can be progressively synthesized to obtain the predicted bit-planes. In the residual bit-plane, the white pixels are the error predictions, and most of them are concentrated in the edges of the image. The error rate of the bit-plane 7 is the lowest because this bit-plane is relatively smooth and just a few edges are presented. The error rate gradually increases as the details of a bit-plane become more complex. Nevertheless, the error rate of bit-plane 1 is still lower than 35% and most bits can be accurately predicted. Therefore, the proposed scheme can progressively predict bit-planes with high accuracy. To further verify the prediction accuracy, Table 1 lists all error rates of different layers in ten test images.


Imagesr7r6r5r4r3r2r1Average

Image 0110.455312.513714.186117.912322.024526.989734.825119.8438
Image 0211.885114.688115.924118.652321.911625.967431.434620.0662
Image 038.451810.781912.576318.415821.481324.804731.776418.3269
Image 0413.220211.450214.526418.191521.569826.765434.069819.9705
Image 059.585610.957313.783319.226126.554924.684133.390819.7403
Image 0610.28149.487913.613922.384622.314526.442033.320619.6921
Image 075.97239.681714.389016.810620.396425.358631.326317.7050
Image 0811.337311.634812.829616.276619.996624.041731.654418.2530
Image 0916.119412.007114.662217.591921.821627.304136.109920.8023
Image 1010.571312.008713.368216.180420.051624.563631.166118.2728
Average10.788011.521113.985918.164221.812325.692132.907419.2673

The error rate monotonously increases from r7 to r1. Two exceptions occur in “Image 04” and “Image 06”: the error rate of r6 is lower than r7. The lowest error rate and highest error rate are 5.9723% and 36.1099%, respectively, in r7 of “Image 07” and r1 of “Image 09.” Generally, the average error rate monotonously increases from high bit-plane to low bit-plane. The average error rate for an entire image is 19.2673%.

The compression ratio is greatly dependent on the error rate and the distribution of error bits. A residual bit-plane with a few concentrated error bits can be compressed efficiently. The compression ratio of the residual bit-planes is plotted in Figure 8. Since the number of error bits increases from r7 to r1, the compression ratio decreases as the bit layer decreases. Even in the exception cases of “Image 04” and “Image 06,” where the error bits of r7 are more than r6, the compression ratio of r7 is also greater due to the concentrated distribution of error bits.

To investigate the effect of block size setting, four different sizes are applied to compress the residual. The compression ratio for different block sizes and bit layers is listed in Table 2. The compression ratio of size 4 is significantly greater than the other sizes for all layers of bit-planes.


Block sizer7r6r5r4r3r2r1

22.12951.81241.51121.28141.13271.04490.9922
42.90952.20591.70531.37311.26141.21371.1426
82.15851.55531.27881.11661.04351.03131.0147
161.43721.13831.07751.02661.00361.00721.0027

Additionally, our proposed residual bit-plane compression technique is also compared with the arithmetic coding technique. As shown in Figure 9, the average compression ratio of our proposed scheme is significantly better in the bit-planes of r7 and r6, while the performance of arithmetic coding is slightly better in the bit-planes of r4 and r3. That is because, in the rear residual bit-planes such as r4 to r1, the errors have no continuous distribution to make our compression technique not efficient. For compression of an entire image, the proposed scheme is significantly better than the arithmetic coding. To further verify the applicability in the medical image domain, the compression ratio for all test images is listed in Table 3. In Table 3, and denote that the sources of compression are residuals and original bit-planes, respectively. Additionally, since the original first bit-plane should be transmitted at the beginning, in two types of bit-plane and , the original first bit-plane is transmitted and compressed by different compression methods. As shown in Table 3, the compression ratio of residual is greater than that of the original version. It implies the devised GAN predictor bank is helpful. In addition, the proposed bit-plane compression technique outperforms the arithmetic coding in the scheme of bit-planewise compression. As shown in the experimental results, the compression ratio is severely affected by the smoothness of residual bit-planes. In medical images, most of the textures are concentrated in the region of interest, which is particularly suitable for the proposed predictor bank to work efficiently. Additionally, the parameters of the pretrained predictor bank are crucial for the sender and receiver, so that the predictor bank needs to be protected to ensure applicability and security.


Image 01Image 02Image 03Image 04Image 05

Proposed1.39231.23521.43471.31801.45661.30871.39701.24531.42721.2650
Arithmetic coding1.37391.00621.34751.00981.42551.00201.36291.00221.37831.0101

Image 06Image 07Image 08Image 09Image 10

Proposed1.40031.24091.46091.27221.46151.29531.39171.23761.46011.2835
Arithmetic coding1.38271.00521.45851.00511.41661.00061.33491.00551.41581.0116

5. Conclusions

In this paper, we introduce a GAN predictor bank and a bit-plane compression technique into the progressive transmission of medical images. Experimental results validate the proposed predictor bank can effectively predict bit-planes under the progressive scheme. Furthermore, the proposed adaptive bit-plane compression is more efficient than the arithmetic coding for the current application. This paper intends to introduce deep learning techniques into the classical progressive image transmission and we hope this paper can give some illuminations for future research. In our future works, we will further introduce and improve more convolutional neural networks into the compression and transmission of medical images.

Data Availability

The data used to support the findings of the study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was partially supported by the JSPS KAKENHI (Grant nos. JP16H06302 and JP18H04120) and JST CREST (Grant nos JPMJCR18A6 and JPMJCR20D3), Japan.

References

  1. H. U. Lemke, “Communication networks for medical image transmission,” Strahlentherapie und Onkologie: Organ der Deutschen Röntgengesellschaft, vol. 169, no. 9, pp. 512–520, 1993. View at: Google Scholar
  2. R. Maani, S. Camorlinga, N. Arnason, and R. Eskicioglu, “A practical fast method for medical imaging transmission based on the DICOM protocol,” Medical Imaging 2010: Advanced PACS-Based Imaging Informatics and Therapeutic Applications, vol. 7628, Article ID 76280M, 2010. View at: Google Scholar
  3. R. Maani, S. Camorlinga, and N. Arnason, “A parallel method to improve medical image transmission,” Journal of Digital Imaging, vol. 25, no. 1, pp. 101–109, 2012. View at: Publisher Site | Google Scholar
  4. K.-H. Tzou, “Progressive image transmission: a review and comparison of techniques,” Optical Engineering, vol. 26, no. 7, Article ID 267581, 1987. View at: Publisher Site | Google Scholar
  5. C.-C. Chang, F.-C. Shiue, and T.-S. Chen, “A new scheme of progressive image transmission based on bit-plane method,” in Proceedings of the Fifth Asia-Pacific Conference on and Fourth Optoelectronics and Communications Conference on Communications, pp. 892–895, Singapore, October 1999. View at: Google Scholar
  6. T.-S. Chen, H.-C. Wu, H.-F. Tsai, M. Hsieh, and S.-F. Chiou, “Progressive transmission of two-dimensional gel electrophoresis image based on context features and bit-plane method,” in Proceedings of the IEEE International Conference on Networking, Sensing and Control, pp. 1241–1246, Taipei, Taiwan, March 2004. View at: Google Scholar
  7. L. Wang and M. Goldberg, “Progressive image transmission using vector quantization on images in pyramid form,” IEEE Transactions on Communications, vol. 37, no. 12, pp. 1339–1349, 1989. View at: Publisher Site | Google Scholar
  8. C.-C. Chang, H.-C. Hsia, and T.-S. Chen, “A progressive image transmission scheme based on block truncation coding,” in The Human Society and the Internet Internet-Related Socio-Economic Issues, pp. 383–397, Berlin, Heidelberg, 2001. View at: Publisher Site | Google Scholar
  9. K.-L. Chung and S.-Y. Tseng, “New progressive image transmission based on quadtree and shading approach with resolution control,” Pattern Recognition Letters, vol. 22, no. 14, pp. 1545–1555, 2001. View at: Publisher Site | Google Scholar
  10. C.-C. Chang, G.-X. Xiao, and T.-S. Chen, “A simple prediction method for progressive image transmission,” in Distributed Multimedia Databases: Techniques and Applications, The Idea Group Publishing, Hershey, PA, USA, 2002, http://www.igi-global.com/chapter/distributed-multimedia-databases/8626. View at: Google Scholar
  11. H. Kim, R. Annavajjala, P. Cosman, and L. Milstein, “Source-channel rate optimization for progressive image transmission over block fading relay channels,” IEEE Transactions on Communications, vol. 58, no. 6, pp. 1631–1642, 2010. View at: Publisher Site | Google Scholar
  12. B. C. Dhara and B. Chanda, “A fast progressive image transmission scheme using block truncation coding by pattern fitting,” Journal of Visual Communication and Image Representation, vol. 23, no. 2, pp. 313–322, 2012. View at: Publisher Site | Google Scholar
  13. L. Wang and M. Goldberg, “Progressive image transmission by transform coefficient residual error quantization,” IEEE Transactions on Communications, vol. 36, no. 1, pp. 75–87, 1988. View at: Publisher Site | Google Scholar
  14. W. J. Hwang and H. Derin, “Multiresolution multiresource progressive image transmission,” IEEE Transactions on Image Processing, vol. 4, no. 8, pp. 1128–1140, 1995. View at: Publisher Site | Google Scholar
  15. Y. Lu, J. Zheng, Y. Jiang, M. Yang, B. Fu, and W. Hou, “Progressive image transmission for medical applications based on wavelet transform with a non-uniform scalar quantization scheme,” in Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 4795–4798, New York, NY, USA, August 2006. View at: Google Scholar
  16. C.-C. Chang and T.-C. Lu, “A wavelet-based progressive digital image transmission scheme,” in Proceedings of the First International Conference on Innovative Computing, Information and Control-Volume I (ICICIC’06), pp. 681–684, Beijing, China, September 2006. View at: Google Scholar
  17. C.-P. Huang and C.-C. Li, “Secure and progressive image transmission through shadows generated by multiwavelet transform,” International Journal of Wavelets, Multiresolution and Information Processing, vol. 6, no. 6, pp. 907–931, 2008. View at: Publisher Site | Google Scholar
  18. R. S. Dilmaghani, A. Ahmadian, M. Ghavami, and A. H. Aghvami, “Progressive medical image transmission and compression,” IEEE Signal Processing Letters, vol. 11, no. 10, pp. 806–809, 2004. View at: Publisher Site | Google Scholar
  19. K.-L. Hung, C.-C. Chang, and I.-C. Lin, “Lossless compression-based progressive image transmission scheme,” The Imaging Science Journal, vol. 52, no. 4, pp. 212–224, 2004. View at: Publisher Site | Google Scholar
  20. Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015. View at: Publisher Site | Google Scholar
  21. R. Neelapu, G. Lavanya Devi, and K. Srinivasa Rao, “Deep learning based conventional neural network architecture for medical image classification,” Traitement du signal, vol. 35, no. 2, pp. 169–182, 2018. View at: Publisher Site | Google Scholar
  22. L. Faes, S. K. Wagner, D. J. Fu et al., “Automated deep learning design for medical image classification by health-care professionals with no coding experience: a feasibility study,” The Lancet Digital Health, vol. 1, no. 5, pp. e232–e242, 2019. View at: Publisher Site | Google Scholar
  23. J. Wagner, Y. Xie, Q. Wu, and Y. Xia, “Medical image classification using synergic deep learning,” Medical Image Analysis, vol. 54, pp. 10–19, 2019. View at: Publisher Site | Google Scholar
  24. H. Chen, D. Ni, J. Qin et al., “Standard plane localization in fetal ultrasound via domain transferred deep neural networks,” IEEE Journal of Biomedical and Health Informatics, vol. 19, no. 5, pp. 1627–1636, 2015. View at: Publisher Site | Google Scholar
  25. Y. Ni, M. Landis, D. T. Laidley, A. Kornecki, A. Lum, and S. Li, “Multi-modal vertebrae recognition using transformed deep convolution network,” Computerized Medical Imaging and Graphics, vol. 51, pp. 11–19, 2016. View at: Publisher Site | Google Scholar
  26. A. Kumar, “Plane identification in fetal ultrasound images using saliency maps and convolutional neural networks,” in Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), pp. 791–794, Prague, Czech Republic, April 2016. View at: Google Scholar
  27. A. Teramoto, H. Fujita, O. Yamamuro, and T. Tamaki, “Automated detection of pulmonary nodules in PET/CT images: ensemble false-positive reduction using a convolutional neural network technique,” Medical Physics, vol. 43, no. 6, pp. 2821–2827, 2016. View at: Publisher Site | Google Scholar
  28. M. J. J. P. Van Grinsven, B. van Ginneken, C. B. Hoyng, T. Theelen, and C. I. Sánchez, “Fast convolutional neural network training using selective data sampling: application to hemorrhage detection in color fundus images,” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1273–1284, 2016. View at: Publisher Site | Google Scholar
  29. J. M. Wolterink, T. Leiner, B. D. de Vos, R. W. van Hamersvelt, M. A. Viergever, and I. Išgum, “Automatic coronary artery calcium scoring in cardiac CT angiography using paired convolutional neural networks,” Medical Image Analysis, vol. 34, pp. 123–136, 2016. View at: Publisher Site | Google Scholar
  30. P. F. Christ, “Automatic liver and lesion segmentation in CT using cascaded fully convolutional neural networks and 3D conditional random fields,” in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 415–423, Athens, Greece, October 2016. View at: Google Scholar
  31. M. Gao, “Segmentation label propagation using deep convolutional neural networks and dense conditional random field,” in Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), pp. 1265–1268, Prague, Czech Republic, April 2016. View at: Google Scholar
  32. K. Kamnitsas, C. Ledig, V. F. J. Newcombe et al., “Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation,” Medical Image Analysis, vol. 36, pp. 61–78, 2017. View at: Publisher Site | Google Scholar
  33. Y. Ledig, I. Kogan, E. Gelbart, O. Geva, and H. Greenspan, “Visualizing and enhancing a deep learning framework using patients age and gender for chest x-ray image retrieval,” in Proceedings of the SPIE on Medical Imaging, p. 978510, San Diego, CA, USA, 2016. View at: Google Scholar
  34. A. Janowczyk, A. Basavanhally, and A. Madabhushi, “Stain normalization using sparse autoencoders (StaNoSA): application to digital pathology,” Computerized Medical Imaging and Graphics, vol. 57, pp. 50–61, 2017. View at: Publisher Site | Google Scholar
  35. P. Kisilev, E. Sason, E. Barkan, and S. Hashoul, “Medical image description using multi-task-loss CNN,” in Deep Learning and Data Labeling for Medical Applications, Springer, Cham, Switzerland, 2016. View at: Google Scholar
  36. I. J. Goodfellow, “Generative adversarial networks,” ArXiv14062661 Cs Stat, 2014, http://arxiv.org/abs/1406.2661. View at: Google Scholar
  37. P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125–1134, Honolulu, HI, USA, July 2017. View at: Google Scholar
  38. C.-C. Chang, “Adversarial learning for invertible steganography,” IEEE Access, vol. 8, pp. 198425–198435, 2020. View at: Publisher Site | Google Scholar
  39. M. Sato, K. Hotta, A. Imanishi, M. Matsuda, and K. Terai, “Segmentation of cell membrane and nucleus by improving Pix2pix,” Biosignals, pp. 216–220, 2018. View at: Google Scholar
  40. M. Mori, T. Fujioka, L. Katsuta et al., “Feasibility of new fat suppression for breast MRI using pix2pix,” Japanese Journal of Radiology, vol. 38, no. 11, pp. 1075–1081, 2020. View at: Publisher Site | Google Scholar

Copyright © 2021 Ching-Chun Chang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views268
Downloads333
Citations

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.