ReflectNet -- A Generative Adversarial Method for Single Image Reflection Suppression
Andreea Birhala, Ionut Mironica
Taking pictures through glass windows almost always produces undesired reflections that degrade the quality of the photo. The ill-posed nature of the reflection removal problem reached the attention of many researchers for more than decades. The main challenge of this problem is the lack of real training data and the necessity of generating realistic synthetic data. In this paper, we proposed a single image reflection removal method based on context understanding modules and adversarial training to efficiently restore the transmission layer without reflection. We also propose a complex data generation model in order to create a large training set with various type of reflections. Our proposed reflection removal method outperforms state-of-the-art methods in terms of PSNR and SSIM on the SIR benchmark dataset.
Validation of image systems simulation technology using a Cornell Box
Zheng Lyu, Krithin Kripakaran, Max Furth
et al.
We describe and experimentally validate an end-to-end simulation of a digital camera. The simulation models the spectral radiance of 3D-scenes, formation of the spectral irradiance by multi-element optics, and conversion of the irradiance to digital values by the image sensor. We quantify the accuracy of the simulation by comparing real and simulated images of a precisely constructed, three-dimensional high dynamic range test scene. Validated end-to-end software simulation of a digital camera can accelerate innovation by reducing many of the time-consuming and expensive steps in designing, building and evaluating image systems.
Ultrasound differential phase contrast using backscattering and the memory effect
Timothy D. Weber, Nikunj Khetan, Ruohui Yang
et al.
We describe a simple and fast technique to perform ultrasound differential phase contrast (DPC) imaging in arbitrarily thick scattering media. Though configured in a reflection geometry, DPC is based on transmission imaging and is a direct analogue of optical differential interference contrast (DIC). DPC exploits the memory effect and works in combination with standard pulse-echo imaging, with no additional hardware or data requirements, enabling complementary phase contrast (in the transverse direction) without any need for intensive numerical computation. We experimentally demonstrate the principle of DPC using tissue phantoms with calibrated speed-of-sound inclusions.
A general framework for reversible data hiding in encrypted images by reserving room before encryption
Ammar Mohammadi
In this paper a general framework to adopt different predictors for reversible data hiding in the encrypted image is presented. We propose innovative predictors that contribute more significantly than conventional ones results in accomplishing more payload. Reserving room before encryption (RRBE) is designated in the proposed scheme making possible to attain high embedding capacity. In RRBE procedure, pre-processing is allowed before image encryption. In our scheme, pre-processing comprises of three main steps: computing prediction-errors, blocking and labeling of the errors. By blocking, we obviate the need for lossless compression to when content-owner is not enthusiastic. Lossless compression is employed in recent state of the art schemes to improve payload. We surpass prior arts exploiting proper predictors, more efficient labeling procedure and blocking of the prediction-errors.
Combined neural network-based intra prediction and transform selection
Thierry Dumas, Franck Galpin, Philippe Bordes
The interactions between different tools added successively to a block-based video codec are critical to its rate-distortion efficiency. In particular, when deep neural network-based intra prediction modes are inserted into a block-based video codec, as the neural network-based prediction function cannot be easily characterized, the adaptation of the transform selection process to the new modes can hardly be performed manually. That is why this paper presents a combined neural network-based intra prediction and transform selection for a block-based video codec. When putting a single neural network-based intra prediction mode and the learned prediction of the selected LFNST pair index into VTM-8.0, -3.71%, -3.17%, and -3.37% of mean BD-rate reduction in all-intra is obtained.
Editorial: Introduction to the Issue on Deep Learning for Image/Video Restoration and Compression
A. Murat Tekalp, Michele Covell, Radu Timofte
et al.
Recent works have shown that learned models can achieve significant performance gains, especially in terms of perceptual quality measures, over traditional methods. Hence, the state of the art in image restoration and compression is getting redefined. This special issue covers the state of the art in learned image/video restoration and compression to promote further progress in innovative architectures and training methods for effective and efficient networks for image/video restoration and compression.
Adaptively Sparse Regularization for Blind Image Restoration
Ningshan Xu
Image quality is the basis of image communication and understanding tasks. Due to the blur and noise effects caused by imaging, transmission and other processes, the image quality is degraded. Blind image restoration is widely used to improve image quality, where the main goal is to faithfully estimate the blur kernel and the latent sharp image. In this study, based on experimental observation and research, an adaptively sparse regularized minimization method is originally proposed. The high-order gradients combine with low-order ones to form a hybrid regularization term, and an adaptive operator derived from the image entropy is introduced to maintain a good convergence. Extensive experiments were conducted on different blur kernels and images. Compared with existing state-of-the-art blind deblurring methods, our method demonstrates superiority on the recovery accuracy.
Optical image decomposition and noise filtering based on Laguerre-Gaussian modes
Jiantao Ma, Dan Wei, Haocheng Yang
et al.
We propose and experimentally demonstrate an efficient image decomposition in the Laguerre-Gaussian (LG) domain. By developing an advanced computing method, the sampling points are much fewer than those in the existing methods, which can significantly improve the calculation efficiency. The beam waist, azimuthal and radial truncation orders of the LG modes are optimized depending on the image information to be restored. In the experiment, we decompose an image by using about 30000 LG modes and realize a high-fidelity reconstruction. Furthermore, we show image noise reduction through LG domain filtering. Our results open a door for LG-mode based image processing.
en
eess.IV, physics.optics
Segmentation of brain tumor on magnetic resonance imaging using a convolutional architecture
Miriam Zulema Jacobo, Jose Mejia
The brain is a complex organ controlling cognitive process and physical functions. Tumors in the brain are accelerated cell growths affecting the normal function and processes in the brain. MRI scans provides detailed images of the body being one of the most common tests to diagnose brain tumors. The process of segmentation of brain tumors from magnetic resonance imaging can provide a valuable guide for diagnosis, treatment planning and prediction of results. Here we consider the problem brain tumor segmentation using a Deep learning architecture for use in tumor segmentation. Although the proposed architecture is simple and computationally easy to train, it is capable of reaching $IoU$ levels of 0.95.
Deep Transfer Learning for Texture Classification in Colorectal Cancer Histology
Srinath Jayachandran, Ashlin Ghosh
Microscopic examination of tissues or histopathology is one of the diagnostic procedures for detecting colorectal cancer. The pathologist involved in such an examination usually identifies tissue type based on texture analysis, especially focusing on tumour-stroma ratio. In this work, we automate the task of tissue classification within colorectal cancer histology samples using deep transfer learning. We use discriminative fine-tuning with one-cycle-policy and apply structure-preserving colour normalization to boost our results. We also provide visual explanations of the deep neural network's decision on texture classification. With achieving state-of-the-art test accuracy of 96.2% we also embark on using deployment friendly architecture called SqueezeNet for memory-limited hardware.
Automated Cardiothoracic Ratio Calculation and Cardiomegaly Detection using Deep Learning Approach
Isarun Chamveha, Treethep Promwiset, Trongtum Tongdee
et al.
We propose an algorithm for calculating the cardiothoracic ratio (CTR) from chest X-ray films. Our approach applies a deep learning model based on U-Net with VGG16 encoder to extract lung and heart masks from chest X-ray images and calculate CTR from the extents of obtained masks. Human radiologists evaluated our CTR measurements, and $76.5\%$ were accepted to be included in medical reports without any need for adjustment. This result translates to a large amount of time and labor saved for radiologists using our automated tools.
Identification and Classification of Phenomena in Multispectral Satellite Imagery Using a New Image Smoother Method and its Applications in Environmental Remote Sensing
M. Kiani
In this paper a new method of image smoothing for satellite imagery and its applications in environmental remote sensing are presented. This method is based on the global gradient minimization over the whole image. With respect to the image discrete identity, the continuous minimization problem is discretized. Using the finite difference numerical method of differentiation, a simple yet efficient 5*5-pixel template is derived. Convolution of the derived template with the image in different bands results in the discrimination of various image elements. This method is extremely fast, besides being highly precise. A case study is presented for the northern Iran, covering parts of the Caspian Sea. Comparison of the method with the usual Laplacian template reveals that it is more capable of distinguishing phenomena in the image.
A Diffractive Neural Network with Weight-Noise-Injection Training
Jiashuo Shi
We propose a diffractive neural network with strong robustness based on Weight Noise Injection training, which achieves accurate and fast optical-based classification while diffraction layers have a certain amount of surface shape error. To the best of our knowledge, it is the first time that using injection weight noise during training to reduce the impact of external interference on deep learning inference results. In the proposed method, the diffractive neural network learns the mapping between the input image and the label in Weight Noise Injection mode, making the network's weight insensitive to modest changes, which improve the network's noise resistance at a lower cost. By comparing the accuracy of the network under different noise, it is verified that the proposed network (SRNN) still maintains a higher accuracy under serious noise.
Efficient Unpaired Image Dehazing with Cyclic Perceptual-Depth Supervision
Chen Liu, Jiaqi Fan, Guosheng Yin
Image dehazing without paired haze-free images is of immense importance, as acquiring paired images often entails significant cost. However, we observe that previous unpaired image dehazing approaches tend to suffer from performance degradation near depth borders, where depth tends to vary abruptly. Hence, we propose to anneal the depth border degradation in unpaired image dehazing with cyclic perceptual-depth supervision. Coupled with the dual-path feature re-using backbones of the generators and discriminators, our model achieves $\mathbf{20.36}$ Peak Signal-to-Noise Ratio (PSNR) on NYU Depth V2 dataset, significantly outperforming its predecessors with reduced Floating Point Operations (FLOPs).
Generative Adversarial Network for Radar Signal Generation
Thomas Truong, Svetlana Yanushkevich
A major obstacle in radar based methods for concealed object detection on humans and seamless integration into security and access control system is the difficulty in collecting high quality radar signal data. Generative adversarial networks (GAN) have shown promise in data generation application in the fields of image and audio processing. As such, this paper proposes the design of a GAN for application in radar signal generation. Data collected using the Finite-Difference Time-Domain (FDTD) method on three concealed object classes (no object, large object, and small object) were used as training data to train a GAN to generate radar signal samples for each class. The proposed GAN generated radar signal data which was indistinguishable from the training data by qualitative human observers.
3D Computational Cannula Fluorescence Microscopy enabled by Artificial Neural Networks
Ruipeng Guo, Zhimeng Pan, Andrew Taibi
et al.
Computational Cannula Microscopy (CCM) is a high-resolution widefield fluorescence imaging approach deep inside tissue, which is minimally invasive. Rather than using conventional lenses, a surgical cannula acts as a lightpipe for both excitation and fluorescence emission, where computational methods are used for image visualization. Here, we enhance CCM with artificial neural networks to enable 3D imaging of cultured neurons and fluorescent beads, the latter inside a volumetric phantom. We experimentally demonstrate transverse resolution of ~6um, field of view ~200um and axial sectioning of ~50um for depths down to ~700um, all achieved with computation time of ~3ms/frame on a laptop computer.
en
eess.IV, physics.optics
Iterative Joint Ptychography-Tomography with Total Variation Regularization
Huibin Chang, Pablo Enfedaque, Stefano Marchesini
In order to determine the 3D structure of a thick sample, researchers have recently combined ptychography (for high resolution) and tomography (for 3D imaging) in a single experiment. 2-step methods are usually adopted for reconstruction, where the ptychography and tomography problems are often solved independently. In this paper, we provide a novel model and ADMM-based algorithm to jointly solve the ptychography-tomography problem iteratively, also employing total variation regularization. The proposed method permits large scan stepsizes for the ptychography experiment, requiring less measurements and being more robust to noise with respect to other strategies, while achieving higher reconstruction quality results.
Zoom in to where it matters: a hierarchical graph based model for mammogram analysis
Hao Du, Jiashi Feng, Mengling Feng
In clinical practice, human radiologists actually review medical images with high resolution monitors and zoom into region of interests (ROIs) for a close-up examination. Inspired by this observation, we propose a hierarchical graph neural network to detect abnormal lesions from medical images by automatically zooming into ROIs. We focus on mammogram analysis for breast cancer diagnosis for this study. Our proposed network consist of two graph attention networks performing two tasks: (1) node classification to predict whether to zoom into next level; (2) graph classification to classify whether a mammogram is normal/benign or malignant. The model is trained and evaluated on INbreast dataset and we obtain comparable AUC with state-of-the-art methods.
Solving RED with Weighted Proximal Methods
Tao Hong, Irad Yavneh, Michael Zibulevsky
REgularization by Denoising (RED) is an attractive framework for solving inverse problems by incorporating state-of-the-art denoising algorithms as the priors. A drawback of this approach is the high computational complexity of denoisers, which dominate the computation time. In this paper, we apply a general framework called weighted proximal methods (WPMs) to solve RED efficiently. We first show that two recently introduced RED solvers (using the fixed point and accelerated proximal gradient methods) are particular cases of WPMs. Then we show by numerical experiments that slightly more sophisticated variants of WPM can lead to reduced run times for RED by requiring a significantly smaller number of calls to the denoiser.
Implementation of encryption on telemedicine
Ulkar Ahmadova, Laman Mammadova, Behnam Kiani Kalejahi
In the era of technology, data security is one of the most important things that both individuals and companies need. Information plays a huge role in our everyday life and keeping it safe should be our number one priority. Nowadays most of the information is transferred via the internet. One of the ways to use it is telemedicine. With the help of telemedicine, people can have an appointment at the doctors without losing their time or money. All of the information about one's health is transferred through the internet but is it that safe? What techniques are used to provide the safety of our confidential information? To guarantee that the information is not changed or that in case it will be stolen no one can still have access to it.