| Name: | Description: | Size: | Format: | |
|---|---|---|---|---|
| 60.18 MB | Adobe PDF |
Authors
Abstract(s)
The medical equipment used to capture retinal fundus images is generally expensive.
With the development of technology and the emergence of smartphones, new portable
screening options have emerged, one of them being the D-Eye device. This and
other similar devices associated with a smartphone, when compared to specialized
equipment, present lower quality in the retinal video captured, yet with sufficient
quality to perform a medical pre-screening. From this, if necessary, individuals can
be referred for specialized screening, in order to obtain a medical diagnosis.
This dissertation contributes to the development of a framework, which is a tool
that allows grouping a set of developed and explored methods, applied to low-quality
retinal videos. Three areas of intervention were defined: the extraction of relevant
regions in video sequences; creating mosaicing images in order to obtain a summary
image of each retinal video; develop of a graphical interface to accommodate the
previous contributions.
To extract the relevant regions from these videos (the retinal zone), two methods
were proposed, one of them is based on more classical image processing approaches
such as thresholds and Hough Circle transform. The other performs the extraction
of the retinal location by applying a neural network, which is one of the methods
reported in the literature with good performance for object detection, the YOLOv4.
The mosaicing process was divided into two stages; in the first stage, the GLAMpoints
neural network was applied to extract relevant points. From these, some
transformations are carried out to have in the same referential the overlap of common
regions of the images. In the second stage, a smoothing process was performed
in the transition between images.
A graphical interface was developed to encompass all the above methods to
facilitate access to and use of them. In addition, other features were implemented,
such as comparing results with ground truth and exporting videos containing only
regions of interest.
Description
Keywords
Convolutional Neural Network Object detection D-Eye Mosaicing Fundus Retinal images
