End-to-end Adversarial Retinal Image Synthesis
End-to-end Adversarial Retinal Image Synthesis
dc.contributor.author | Costa,P | en |
dc.contributor.author | Adrian Galdran | en |
dc.contributor.author | Maria Inês Meyer | en |
dc.contributor.author | Niemeijer,M | en |
dc.contributor.author | Abramoff,M | en |
dc.contributor.author | Ana Maria Mendonça | en |
dc.contributor.author | Aurélio Campilho | en |
dc.contributor.other | 6071 | en |
dc.contributor.other | 6381 | en |
dc.contributor.other | 6825 | en |
dc.contributor.other | 6835 | en |
dc.date.accessioned | 2019-03-04T15:08:18Z | |
dc.date.available | 2019-03-04T15:08:18Z | |
dc.date.issued | 2017 | en |
dc.description.abstract | In medical image analysis applications, the availability of large amounts of annotated data is becoming increasingly critical. However, annotated medical data is often scarce and costly to obtain. In this paper, we address the problem of synthesizing retinal color images by applying recent techniques based on adversarial learning. In this setting, a generative model is trained to maximize a loss function provided by a second model attempting to classify its output into real or synthetic. In particular, we propose to implement an adversarial autoencoder for the task of retinal vessel network synthesis. We use the generated vessel trees as an intermediate stage for the generation of color retinal images, which is accomplished with a Generative Adversarial Network. Both models require the optimization of almost everywhere differentiable loss functions, which allows us to train them jointly. The resulting model offers an end-to-end retinal image synthesis system capable of generating as many retinal images as the user requires, with their corresponding vessel networks, by sampling from a simple probability distribution that we impose to the associated latent space. We show that the learned latent space contains a well-defined semantic structure, implying that we can perform calculations in the space of retinal images, e.g., smoothly interpolating new data points between two retinal images. Visual and quantitative results demonstrate that the synthesized images are substantially different from those in the training set, while being also anatomically consistent and displaying a reasonable visual quality. IEEE | en |
dc.identifier.uri | http://repositorio.inesctec.pt/handle/123456789/8302 | |
dc.identifier.uri | http://dx.doi.org/10.1109/tmi.2017.2759102 | en |
dc.language | eng | en |
dc.rights | info:eu-repo/semantics/openAccess | en |
dc.title | End-to-end Adversarial Retinal Image Synthesis | en |
dc.type | article | en |
dc.type | Publication | en |
Files
Original bundle
1 - 1 of 1