RAG-FW: A Hybrid Convolutional Framework for the Automated Extraction of Retinal Lesions and Lesion-Influenced Grading of Human Retinal Pathology
Overview
Medical Informatics
Affiliations
The identification of retinal lesions plays a vital role in accurately classifying and grading retinopathy. Many researchers have presented studies on optical coherence tomography (OCT) based retinal image analysis over the past. However, to the best of our knowledge, there is no framework yet available that can extract retinal lesions from multi-vendor OCT scans and utilize them for the intuitive severity grading of the human retina. To cater this lack, we propose a deep retinal analysis and grading framework (RAG-FW). RAG-FW is a hybrid convolutional framework that extracts multiple retinal lesions from OCT scans and utilizes them for lesion-influenced grading of retinopathy as per the clinical standards. RAG-FW has been rigorously tested on 43,613 scans from five highly complex publicly available datasets, containing multi-vendor scans, where it achieved the mean intersection-over-union score of 0.8055 for extracting the retinal lesions and the accuracy of 98.70% for the correct severity grading of retinopathy.
OCT5k: A dataset of multi-disease and multi-graded annotations for retinal layers.
Arikan M, Willoughby J, Ongun S, Sallo F, Montesel A, Ahmed H Sci Data. 2025; 12(1):267.
PMID: 39952954 PMC: 11829038. DOI: 10.1038/s41597-024-04259-z.
Azizi M, Abhari S, Sajedi H PLoS One. 2024; 19(6):e0304943.
PMID: 38837967 PMC: 11152295. DOI: 10.1371/journal.pone.0304943.
Krzywicki T, Brona P, Zbrzezny A, Grzybowski A J Clin Med. 2023; 12(10).
PMID: 37240693 PMC: 10219065. DOI: 10.3390/jcm12103587.
Ai Z, Huang X, Feng J, Wang H, Tao Y, Zeng F Front Neuroinform. 2022; 16:876927.
PMID: 35784186 PMC: 9243322. DOI: 10.3389/fninf.2022.876927.
Continual Learning Objective for Analyzing Complex Knowledge Representations.
Khan A, Hassan T, Akram M, Alghamdi N, Werghi N Sensors (Basel). 2022; 22(4).
PMID: 35214568 PMC: 8879446. DOI: 10.3390/s22041667.