https://eugs.org
Policy-Driven, Multimodal Deep Learning for Predicting Visual Fields from the Optic Disc and Optical Coherence Tomography Imaging
March 31, 2022

Policy-Driven, Multimodal Deep Learning for Predicting Visual Fields from the Optic Disc and Optical Coherence Tomography Imaging

Author(s): Yuka Kihara (1), Giovanni Montesano (2), Andrew Chen (1), Nishani Amerasinghe (3), Chrysostomos Dimitriou (4), Aby Jacob (3), Almira Chabi (5), David P Crabb (2), Aaron Y Lee (6)
1 University of Washington, Department of Ophthalmology, Seattle, WA, USA.
2 City, University of London, Optometry and Visual Sciences, London, UK.
3 University Hospital Southampton NHS Foundation Trust, Southampton, UK.
4 Colchester Hospital, East Suffolk and North Essex NHS Foundation Trust, Colchester, UK.
5 Santen, Emeryville, CA, USA.
6 University of Washington, Department of Ophthalmology, Seattle, WA, USA. Electronic address: aaronylee@gmail.com.

PURPOSE: To develop and validate a deep learning (DL) system for predicting each point on visual fields (VF) from disc and optical coherence tomography (OCT) imaging and derive a structure-function mapping.

DESIGN: Retrospective, cross-sectional database study PARTICIPANTS: 6437 patients undergoing routine care for glaucoma in three clinical sites in the UK.

METHODS: OCT and infrared reflectance (IR) optic disc imaging was paired with the closest VF within 7 days. Efficient-Net B2 was used to train two single modality DL models to predict each of the 52 sensitivity points on the 24-2 VF pattern. A policy DL model was designed and trained to fuse the two model predictions.

MAIN OUTCOME MEASURES: Pointwise Mean Absolute Error (PMAE) RESULTS: A total of 5078 imaging to VF pairs were used as a held-out test set to measure the final performance. The improvement in PMAE with the policy model was 0.485 0.533 dB compared to the IR image of the disc alone and 0.060 0.073 dB compared to the OCT alone. The improvement with the policy fusion model was statistically significant (p < 0.0001). Occlusion masking shows that the DL models learned the correct structure function mapping in a data-driven, feature agnostic fashion.

CONCLUSIONS: The multimodal, policy DL model performed the best; it provided explainable maps of its confidence in fusing data from single modalities and provides a pathway for probing the structure-function relationship in glaucoma.

Copyright © 2022. Published by Elsevier Inc.

Ophthalmology. 2022 Feb 21;S0161-6420(22)00156-7. doi: 10.1016/j.ophtha.2022.02.017. Online ahead of print.

PMID: 35202616 DOI: 10.1016/j.ophtha.2022.02.017

Keywords: Artificial intelligence; Deep learning; Glaucoma; Optical coherence tomography; Perimetry; Structure-function; Visual field


Tip of the Month manager: Anthony Khawaja
Tip of the Month editorial board: Humma Shahid, Karl Mercieca, Francisco Goni
Tip of the Month editors in chief: Francesco Oddone, Manuele Michelessi