A Single-Parameter Factor-Graph Image Prior
Tianyang Wang, Ender Konukoglu, Hans-Andrea Loeliger
We propose a novel piecewise smooth image model with piecewise constant local parameters that are automatically adapted to each image. Technically, the model is formulated in terms of factor graphs with NUP (normal with unknown parameters) priors, and the pertinent computations amount to iterations of conjugate-gradient steps and Gaussian message passing. The proposed model and algorithms are demonstrated with applications to denoising and contrast enhancement.
Lung Cancer detection using Deep Learning
Aryan Chaudhari, Ankush Singh, Sanchi Gajbhiye
et al.
In this paper we discuss lung cancer detection using hybrid model of Convolutional-Neural-Networks (CNNs) and Support-Vector-Machines-(SVMs) in order to gain early detection of tumors, benign or malignant. The work uses this hybrid model by training upon the Computed Tomography scans (CT scans) as dataset. Using deep learning for detecting lung cancer early is a cutting-edge method.
A square cross-section FOV rotational CL (SC-CL) and its analytical reconstruction method
Xiang Zou, Wuliang Shi, Muge Du
et al.
Rotational computed laminography (CL) has broad application potential in three-dimensional imaging of plate-like objects, as it only needs x-ray to pass through the tested object in the thickness direction during the imaging process. In this study, a square cross-section FOV rotational CL (SC-CL) was proposed. Then, the FDK-type analytical reconstruction algorithm applicable to the SC-CL was derived. On this basis, the proposed method was validated through numerical experiments.
Fast probabilistic snake algorithm
Jérôme Gilles, Bertrand Collin
Few people use the probability theory in order to achieve image segmentation with snake models. In this article, we are presenting an active contour algorithm based on a probability approach inspired by A. Blake work and P. R{é}fr{é}gier's team research in France. Our algorithm, both very fast and highly accurate as far as contour description is concerned, is easily adaptable to any specific application.
An adaptive Lagrange multiplier determination method for rate-distortion optimisation in hybrid video codecs
Fan Zhang, David R. Bull
This paper describes an adaptive Lagrange multiplier determination method for rate-quality optimisation in video compression. Inspired by the experimental results of a Lagrange multiplier selection test, the presented approach adaptively estimates the optimum Lagrange multiplier for different video content, based on distortion statistics of recently encoded frames. The proposed algorithm has been fully integrated into both the H.264 and HEVC reference codecs, and is used in rate-distortion optimisation for encoding B frames. The results show promising (up to 11% on the sequences tested) overall bitrate savings, for a minimal increase in complexity, on various types of test content based on Bjontegaard delta measurements.
Big Plastic Masses Detection using Sentinel 2 Images
Fernando Martin-Rodriguez
This communication describes a preliminary research on detection of big masses of plastic (marine litter) on the oceans and seas using EO (Earth Observation) satellite systems. Free images from the Sentinel 2 (Copernicus Project) platform are used. To develop a plastic recognizer, we start with an image where we can find a big accumulation of "nonfloating" plastic: Almería greenhouses. We made a test using remote sensing differential indexes, but we got much better results using all available wavelengths (thirteen frequency bands) and applying Neural Networks to that feature vector.
Beyond multi-view deconvolution for inherently-aligned fluorescence tomography
Daniele Ancora, Gianluca Valentini, Antonio Pifferi
et al.
In multi-view fluorescence microscopy, each angular acquisition needs to be aligned with care to obtain an optimal volumetric reconstruction. Here, instead, we propose a neat protocol based on auto-correlation inversion, that leads directly to the formation of inherently aligned tomographies. Our method generates sharp reconstructions, with the same accuracy reachable after sub-pixel alignment but with improved point-spread-function. The procedure can be performed simultaneously with deconvolution further increasing the reconstruction resolution.
en
eess.IV, physics.comp-ph
Experimental digital Gabor hologram rendering by a model-trained convolutional neural network
J. Rivet, A. Taliercio, C. Fang
et al.
Digital hologram rendering can be performed by a convolutional neural network, trained with image pairs calculated by numerical wave propagation from sparse generating images. 512-by-512 pixeldigital Gabor magnitude holograms are successfully estimated from experimental interferograms by a standard UNet trained with 50,000 synthetic image pairs over 70 epochs.
en
eess.IV, physics.data-an
100,000 frames-per-second compressive imaging with a conventional rolling-shutter camera by random point-spread-function engineering
Gil Weinberg, Ori Katz
We demonstrate an approach that allows taking videos at very high-speeds of over 100,000 frames per second (fps) by exploiting the fast sampling rate of the standard rolling-shutter readout mechanism, common to most conventional sensors, and a compressive-sampling acquisition scheme. Our approach is directly applied to a conventional imaging system by the simple addition of a diffuser to the pupil plane, randomly encoding the entire field-of-view to each camera row, while maintaining diffraction-limited resolution. A short video is reconstructed from a single camera frame via a compressed-sensing reconstruction algorithm, exploiting inherent sparsity of the imaged scene.
en
eess.IV, physics.optics
MXR-U-Nets for Real Time Hyperspectral Reconstruction
Atmadeep Banerjee, Akash Palrecha
In recent times, CNNs have made significant contributions to applications in image generation, super-resolution and style transfer. In this paper, we build upon the work of Howard and Gugger, He et al. and Misra, D. and propose a CNN architecture that accurately reconstructs hyperspectral images from their RGB counterparts. We also propose a much shallower version of our best model with a 10% relative memory footprint and 3x faster inference, thus enabling real-time video applications while still experiencing only about a 0.5% decrease in performance.
Uncertainty Evaluation Metric for Brain Tumour Segmentation
Raghav Mehta, Angelos Filos, Yarin Gal
et al.
In this paper, we develop a metric designed to assess and rank uncertainty measures for the task of brain tumour sub-tissue segmentation in the BraTS 2019 sub-challenge on uncertainty quantification. The metric is designed to: (1) reward uncertainty measures where high confidence is assigned to correct assertions, and where incorrect assertions are assigned low confidence and (2) penalize measures that have higher percentages of under-confident correct assertions. Here, the workings of the components of the metric are explored based on a number of popular uncertainty measures evaluated on the BraTS 2019 dataset.
Accelerating computed tomographic imaging spectrometer reconstruction using a parallel algorithm exploiting spatial shift-invariance
Larz White, W. Bryan Bell, Ryan Haygood
Computed Tomographic Imaging Spectrometers (CTIS) capture hyperspectral images in realtime. However, post processing the imagery can require enormous computational resources; thus, limiting its application to non-realtime scenarios. To overcome these challenges we developed a highly parallelizable algorithm that exploits spatial shift-invariance. To demonstrate the versatility of our new algorithm, we developed implementations on a desktop and an embedded graphics processing unit (GPU). To our knowledge, our results show the fastest image reconstruction times reported.
Exploiting Temporal Attention Features for Effective Denoising in Videos
Aryansh Omray, Samyak Jain, Utsav Krishnan
et al.
Video Denoising is one of the fundamental tasks of any videoprocessing pipeline. It is different from image denoising due to the tem-poral aspects of video frames, and any image denoising approach appliedto videos will result in flickering. The proposed method makes use oftemporal as well as spatial dimensions of video frames as part of a two-stage pipeline. Each stage in the architecture named as Spatio-TemporalNetwork uses a channel-wise attention mechanism to forward the encodersignal to the decoder side. The Attention Block used in this paper usessoft attention to ranks the filters for better training.
A SARS-CoV-2 Microscopic Image Dataset with Ground Truth Images and Visual Features
Chen Li, Jiawei Zhang, Frank Kulwa
et al.
SARS-CoV-2 has characteristics of wide contagion and quick propagation velocity. To analyse the visual information of it, we build a SARS-CoV-2 Microscopic Image Dataset (SC2-MID) with 48 electron microscopic images and also prepare their ground truth images. Furthermore, we extract multiple classical features and novel deep learning features to describe the visual information of SARS-CoV-2. Finally, it is proved that the visual features of the SARS-CoV-2 images which are observed under the electron microscopic can be extracted and analysed.
GANs 'N Lungs: improving pneumonia prediction
Tatiana Malygina, Elena Ericheva, Ivan Drokin
We propose a novel method to improve deep learning model performance on highly-imbalanced tasks. The proposed method is based on CycleGAN to achieve balanced dataset. We show that data augmentation with GAN helps to improve accuracy of pneumonia binary classification task even if the generative network was trained on the same training dataset.
Deep Learning Based Computed Tomography Whys and Wherefores
Shabab Bazrafkan, Vincent Van Nieuwenhove, Joris Soons
et al.
This is an article about the Computed Tomography (CT) and how Deep Learning influences CT reconstruction pipeline, especially in low dose scenarios.
Use of convexity in contour detection
Victor Churchill
In this paper, we formulate a simple algorithm that detects contours around a region of interest in an image. After an initial smoothing, the method is based on viewing an image as a topographic surface and finding convex and/or concave regions using simple calculus-based testing. The algorithm can achieve multi-scale contour detection by altering the initial smoothing. We show that the method has promise by comparing results on several images with the watershed transform performed on the gradient images.
Attention Based Image Compression Post-Processing Convolutional Neural Network
Yuyang Xue, Jiannan Su
The traditional image compressors, e.g., BPG and H.266, have achieved great image and video compression quality. Recently, Convolutional Neural Network has been used widely in image compression. We proposed an attention-based convolutional neural network for low bit-rate compression to post-process the output of traditional image compression decoder. Across the experimental results on validation sets, the post-processing module trained by MAE and MS-SSIM losses yields the highest PSNR of 32.10 on average at the bit-rate of 0.15.
Efficient Nonlinear Transforms for Lossy Image Compression
Johannes Ballé
We assess the performance of two techniques in the context of nonlinear transform coding with artificial neural networks, Sadam and GDN. Both techniques have been successfully used in state-of-the-art image compression methods, but their performance has not been individually assessed to this point. Together, the techniques stabilize the training procedure of nonlinear image transforms and increase their capacity to approximate the (unknown) rate-distortion optimal transform functions. Besides comparing their performance to established alternatives, we detail the implementation of both methods and provide open-source code along with the paper.
Single molecule localization by $\ell_2-\ell_0$ constrained optimization
Arne Bechensteen, Laure Blanc-Féraud, Gilles Aubert
Single Molecule Localization Microscopy (SMLM) enables the acquisition of high-resolution images by alternating between activation of a sparse subset of fluorescent molecules present in a sample and localization. In this work, the localization problem is formulated as a constrained sparse approximation problem which is resolved by rewriting the $\ell_0$ pseudo-norm using an auxiliary term. In the preliminary experiments with the simulated ISBI datasets the algorithm yields as good results as the state-of-the-art in high-density molecule localization algorithms.