Abstract
Spaceflight associated neuro-ocular syndrome (SANS) is one of the largest physiologic barriers to spaceflight and requires evaluation and mitigation for future planetary missions. As the spaceflight environment is a clinically limited environment, the purpose of this research is to provide automated, early detection and prognosis of SANS with a machine learning model trained and validated on astronaut SANS optical coherence tomography (OCT) images. In this study, we present a lightweight convolutional neural network (CNN) incorporating an EfficientNet encoder for detecting SANS from OCT images titled “SANS-CNN.” We used 6303 OCT B-scan images for training/validation (80%/20% split) and 945 for testing with a combination of terrestrial images and astronaut SANS images for both testing and validation. SANS-CNN was validated with SANS images labeled by NASA to evaluate accuracy, specificity, and sensitivity. To evaluate real-world outcomes, two state-of-the-art pre-trained architectures were also employed on this dataset. We use GRAD-CAM to visualize activation maps of intermediate layers to test the interpretability of SANS-CNN’s prediction. SANS-CNN achieved 84.2% accuracy on the test set with an 85.6% specificity, 82.8% sensitivity, and 84.1% F1-score. Moreover, SANS-CNN outperforms two other state-of-the-art pre-trained architectures, ResNet50-v2 and MobileNet-v2, in accuracy by 21.4% and 13.1%, respectively. We also apply two class-activation map techniques to visualize critical SANS features perceived by the model. SANS-CNN represents a CNN model trained and validated with real astronaut OCT images, enabling fast and efficient prediction of SANS-like conditions for spaceflight missions beyond Earth’s orbit in which clinical and computational resources are extremely limited.
Similar content being viewed by others
Introduction
Spaceflight-associated neuro-ocular syndrome (SANS) refers to a unique constellation of neuro-ophthalmic imaging and clinical findings observed in astronauts after long duration spaceflight (LDSF)1,2. These findings include unilateral/asymmetric/bilateral optic disc edema of different Frisén grades, chorioretinal folds, posterior globe flattening, hyperopic refractive error shift, and cotton wool spots3. Although terrestrial analogs have been helpful in understanding SANS, there is no terrestrial equivalent and ongoing research is being conducted to further understand the underlying pathogenesis of this neuro-ophthalmic phenomenon4,5,6,7,8.
Assigned by the National Aeronautics and Space Administration (NASA), SANS has an elevated Likelihood and Consequence Ratio for a planetary mission to Mars, indicating potential long-term health consequences to astronauts, thus serving as one of the largest physiologic barriers to future spaceflight9,10,11. Mitigation research include lower body negative pressure12,13, pressurized goggles14, nutrition15, and understanding genetic predispositions to developing SANS8. While countermeasure efforts are promising, anticipated missions that expose astronauts to microgravity longer than what is known continue to require critical analysis of SANS progression. Moreover, anticipatory guidance for employing such countermeasures for SANS requires timely evaluation during deep-space exploration. On board the International Space Station (ISS), imaging modalities including optical coherence tomography (OCT), fundus imaging, and orbital ultrasound have been instrumental in providing high-quality analysis of SANS progression3,16,17,18. On deep space missions beyond Earth’s orbit, such as the mission to Mars and beyond, critically analyzing these images will be particularly challenging. Additionally, increases in transmission latency for these large data images to Earth may be a large barrier for an expert reader on Earth to critically analyze SANS progression in a timely fashion, particularly as spacecraft continues to travel farther from Earth19,20,21. This isolation from an expert evaluator of SANS progression during these missions may be mitigated with an automated, expert system that can analyze imaging data onboard the spacecraft. A primary consideration of such a system in the spaceflight environment is the limited computational capacity.
In this paper, we report the development, validation, and accuracy of a deep learning architecture with lightweight Convolutional Neural Networks (CNNs) designed specifically to detect SANS findings on OCT B-scans for future deep space exploration entitled “SANS-CNN”. We utilize SANS images provided and labeled by NASA to validate this architecture. We compare SANS-CNN to state-of-the-art, publicly available pre-trained architectures to interpret the robustness of such a system for spaceflight. The two primary aims of this architecture are: (1) to provide an accurate model that properly detects SANS-like features and (2) to utilize efficient and lightweight CNN predictions for the computationally limited spaceflight environment. The development of technology for detecting and evaluating SANS progression may serve to provide astronauts immediate guidance on whether escalation of SANS countermeasures (e.g., lower body pressure) is warranted during deep-space missions.
Methodology
Pre-trained Encoder
Utilizing pre-trained models for transfer learning tasks has shown tremendous promise in healthcare22,23,24, physics-driven simulation25, drug-discovery26, and computational biology27. Pre-trained models are architectures previously trained on a large data set, and then the weights of these models are transferred and trained on a downstream task. For example, a popular deep convolutional neural network called ResNet28 has been trained using 3.2 million ImageNet29 and images have been used for many downstream tasks for ophthalmology. Similar architectures with residual blocks have seen many adoptions in challenging downstream tasks for ophthalmology such as image-to-image translation, image inpainting, and image segmentation30,31. In Fig. 1, our proposed Efficient and Interpretable Deep ConvNet architecture is given, which consists of an encoder which takes the OCT B-Scans as input, a decoder, and an output for predicting between SANS and Normal cases. The encoder consists of a pre-trained network with multiple residual and downsampling blocks, as shown in Fig. 2.
The basic structure of ResNet consists of a residual unit with two consecutive convolution layers and a skip connection that adds the feature tensor of the input with the output28. However, the authors improved upon this, utilized pre-activation with batch-normalization to address the vanishing gradient problem, and proposed a new architecture called ResNet-v2 with new residual and downsampling blocks. In terms of efficiency, two other architectures have proposed modified learnable blocks which utilize fewer parameters: MobileNet32 and EfficientNet33. Similar to ResNet, the authors of both these models improved upon their architecture and proposed two new architectures, MobileNetV232 and EfficientNetV233. For our experimentation, we use pre-trained encoders of these three architectures. Moreover, all of them were trained on ImageNet2012 datasets29. The difference and similarities of these three architectures are given in Table 1.
Building Blocks
First, we use the ResNet50-v2 architecture for our pre-trained encoder. The encoder consists of residual downsampling and residual identity blocks with pre-activation. We illustrate both of these residual blocks in Fig. 3 (I) and (ii). The residual downsampling blocks have three sub-blocks successively with Batch-Normalization, ReLU, and Convolution. Also, a skip connection with the convolution layer is added from the first sub-blocks ReLU with the last convolution layer’s output. The first and last convolution has kernel size, k = 3 and stride, s = 1. The second and skip-connection convolutions have stride, s = 2. Kernel size and stride are utilized to determine the receptive field of the convolution operation to extract intrinsic spatial information from the images. We use a small kernel size and stride size to extract fine local features, which helps with better downstream information aggregation. On the other hand, downsampling layers are employed to decrease the image features spatially (typically half of the original by using stride = 2) for the convolution operation to work on, which helps with utilizing less memory. By combining small kernel and stride size and using downsampling after each stage, we minimize the overall number of parameters, which effectively helps with overall memory utilization and redundant feature extraction.
Next, we use the MobileNetV2 pre-trained encoder, consisting of downsampling and identity blocks. Both of these blocks are visualized in Fig. 3 (iii) and (iv). Unlike ResNetV2, the MobileNetV2 uses post-activation. Both downsampling and identity blocks consist of three sub-blocks. The first and last sub-blocks have Convolution, Batch-normalization, and ReLU activation layers. The second sub-block has Depth-wise Convolution, Batch-normalization, and ReLU activation layers. The identity block has a skip connection from the input and is added with the output of the last ReLU layer. The first and last convolution has kernel size, k = 3 and stride, s = 1. The depth-wise convolution has stride, s = 2.
Lastly, we incorporate the EfficientNetV2 pre-trained encoder, which consists of Squeeze-and-excitation blocks. The block is illustrated in Fig. 3(v). The block consists of three sub-blocks, with the first and third consisting of Convolution, Batch-Normalization, and Swish activation layers. The second sub-block consists of Global-average pooling, Convolution, Swish activation, Convolution, and Softmax activation layers. The output of the second sub-block is element-wise multiplied by the first sub-block’s output. The convolution in the sub-block has a kernel size, k = 3, and the rest of the convolution in other sub-blocks has a kernel size, k = 1. As Efficient-NetV2 only utilizes this block for downsampling and regular blocks, the stride size is changed in the first sub-block depending on the block definition. The stride size is chosen as, s = 2 for downsampling and s = 1 for a regular block.
Decoder
The decoder consists of a global average pooling layer, followed by three dense layers as illustrated in Fig. 2. The global average pooling takes the average of the spatial dimensions and transforms it into a 1D feature vector. The three dense layers consist of 256, 64, and 2 neurons. The first two dense layers are followed by a dropout layer which randomly drops activations to zero. We use drop rate of 0.6 and 0.2 (out of 1) for the dropout layers successively. We use two neurons in the last layer for predicting SANS or Normal class. The downstream task is a supervised-classification task, as the ground-truth values of one hot-encoded vector.
Objective function
For training we incorporate weighted categorical cross-entropy loss as shown in Equation (1). Here, \({y}_{i}\) is the ground-truth class and \({y}_{i}^{{\prime} }\) is the predicted class. The \(c\) signifies the number of samples, \({w}_{i}\) signifies weight values, and \({cat}\) denotes that is a categorical loss. The weight values were chosen based on the number of samples per-class.
Equation 1. Weighted categorical cross-entropy loss.
Data preprocessing
For our experimentation, we use the OCT-B scans that are separated for training, validation and testing based on unique astronauts. The control images are from pre-flight, in-flight and post-flight OCT volumes. The SANS images were taken from both in-flight and post-flight OCT volumes. For training, we utilized 2506 SANS and 3797 Normal OCT B-scans. For validation, we incorporated 627 SANS and 950 Normal OCT B-scans. In order to evaluate on a hold-out test set, we utilized 467 SANS and 478 Normal images. All images were normalized to have values between 0–1 from 0–255-pixel intensities for training, validation and testing and the images were pre-processed to have a resolution size of 512 × 512.
Hyper-parameter Tuning
For training all the models, we used Adam optimizer for adapting the parameter learning rates34. The initial learning rate, we used the Adam optimizer with an initial learning rate of \({\rm{lr}}=0.0001.\) We utilized a mini-batch of \(b=16\) and trained all methods for 30 epochs. For the class weights, we chose 0.83 for majority (Normal) and 1.257 for minority (SANS) training samples. We used the Keras Deep learning library (https://keras.io/) with Tensorflow backend (https://www.tensorflow.org/) to train our models. For training, we utilized two callbacks: “Reduce Learning rate on Plateau” and “Model Checkpointer”. The first callback reduced the learning rate by \(0.1\) if the validation loss did not decrease for six epochs. Whereas the other callback saves the best snapshot of the model weights for each epoch.
Metrics
We used three standard metrics for calculating the Accuracy, Sensitivity (True Positive Rate), Specificity (True Negative Rate), Precision (Positive Predictive Value) and F1-score. The metrics are calculated as follows in Equation (2):
Equation 2. Equations to calculate accuracy, sensitivity, specificity, precision, and F1-score. Here, N is the number of samples, and K is the number of classes (K = 2). TP= True Positive, FP = False Positive, FN= False Negative, TN = True Negative.
Results
Quantitative results
As seen in Table 2, the best-performing model, EfficientNet-v2, achieved 84.2% accuracy, 85.6% specificity, 82.8% sensitivity, and an 84.1% F1-score. Compared to that, ResNet50-v2 achieved differences of 17–23% less across all six metrics. In contrast, MobileNet-v2 achieved a difference of 12.1% improvement in sensitivity but got lower precision, specificity, F1-score and accuracy. It is evident that EfficientNet-v2 has better overall performance in this SANS vs. Normal images recognition task.
Qualitative results
For producing “visual justifications” for decisions made by our ConvNet models for accurate classification, we use GRAD-CAM35, and GRAD-CAM++36. These techniques utilize back-propagated gradients to show important regions of the images from the perspective of a specific layer, intensified for the classification decision’s maximum probability. In Fig. 3, we visualize the differences in activations of three blocks of the encoder layers for three of our models on a SANS image. Stage N is the last convolution blocks’ output before the global average pooling layer. Similarly, Stage N-1 and Stage N-2 are the previous convolution block’s outputs. From Fig. 1, row 1, it is apparent that MobileNetV2 had sparsely activated signals in choroidal fold region. However, in row 2, ResNet50v2 had concentrated activated signals and activations in the choroidal folds and the unimportant region of Bruch’s Membrane (BM) and retinal pigment epithelium (RPE). In contrast, our best architecture, EfficientNetV2, had the best output for choroidal folds manifested in the RPE and BM as well as the cotton wool spots manifesting below the retinal nerve fiber layer (RNFL). This visualization conforms to the patterns seen in the output metrics where EfficientNevV2 performs the best in most metrics while ResNet50v2 is the lowest performing. This qualitative illustration compliments our model’s overall explainability and knowledge transferability.
Discussion
The results from SANS-CNN demonstrate a robust detection of SANS findings with lightweight CNNs compared to state-of-the-art pre-trained architectures. Given that prolonged optic disc edema and chorioretinal folds have been observed to lead to potential significant visual impairment terrestrially (e.g., chorioretinal folds-related maculopathy or idiopathic intracranial hypertension)37,38,39,40, the automatic evaluation of SANS for immediate insight to onboard members is critical for deep space exploration. As mitigation strategies continue to develop for SANS, there is a necessity for guidance on how long astronauts should employ these countermeasures during spaceflight. Countermeasures such as LBNP and goggles have shown promise for counteracting SANS13,14. LBNP attenuates the cephalad fluid shifts during spaceflight that result from a reduction in hydrostatic pressure in microgravity. These cephalad fluid shifts have been hypothesized to be a critical component to SANS development. In a randomized crossover trial in a terrestrial setting, Hearon et al. observed that LBNP at night mitigated an increase in choroidal area and volume measured by OCT during supine bed rest, suggesting that this countermeasure during sleep on spaceflight missions may counteract SANS. During prolonged exposure to microgravity than what is known, SANS may continue to progress despite nightly LBNP. Automated and timely evaluation of SANS may provide guidance to whether further escalation of lower body negative pressure beyond just nightly use is necessary to attenuate SANS progression.
The utilization of SANS-CNN can be further validated with terrestrial analogs of SANS. Head-down tilt bed rest is a promising terrestrial analog of SANS which mimics cephalad fluid shifts in spaceflight and has been observed to produce optic disc edema and chorioretinal folds41. In conjunction with countermeasures during head-down tilt bed rest, SANS-CNN may help to provide an objective evaluation of SANS-like progression with terrestrial analogs.
Several limitations exist for this study that may be considered for future technical development. First, the labelled OCT data provided by NASA spans across many years with different evaluating experts which limits the uniformity of data. This can be addressed with prospective studies in SANS with uniform labelers. Secondly, given the natural limitation of humans that go to space and the relatively recent discovery of SANS, the number of SANS images may be increased in the future to provide a more robust training dataset. This limitation can be approached with deep learning transfer learning techniques that can build upon larger datasets with terrestrial data to strengthen a smaller training dataset42,43. However, SANS has no terrestrial equivalent; thus, utilizing astronaut SANS images from future spaceflights will be of utmost importance. Though we achieved over 80–85% on five different metrics, there is room for improvement to enhance the architecture to achieve accuracies above 95%. Given, the model is receiving OCT B-scans image inputs for healthy and SANS patients, some B-scans from the SANS patients might not contain apparent SANS indicator (like choroidal folds or cotton wool spots), which might explain the model’s score not reaching 90% or beyond. In the future, our team seeks to investigate whether pre-training on non-SANS publicly available retinal degenerative images (e.g., age-related macular degeneration and diabetic macular edema) can improve the overall performance of our model. Additionally, we seek to explore the incorporation of Generative Adversarial Networks to generate synthetic SANS data to improve the model further20,44,45. Lastly, in the extraterrestrial environment with limited resources such as an exploratory mission traveling away from Earth, optimizing computational utilization and lightweight aspects of such an automated system is critical for systems resourcefulness. Thus, continuing to optimize the computational efficiency of this system is of utmost importance.
Automated detection of SANS during deep-space exploration where clinical resources are limited may provide necessary and timely guidance on countermeasure utilization. SANS-CNN is designed specifically for astronaut ophthalmic health and demonstrates robust detection of SANS-like conditions when compared to existing state-of-the-art architectures. This work is a part of a comprehensive ocular framework to monitor SANS with deep learning and extended reality46,47,48,49,50,51. Future directions include developing and merging models with other modalities onboard the ISS such as fundus photography, which is useful for detecting optic disc edema, chorioretinal folds, and cotton wool spots. As OCTA information becomes available, labelled data with SANS can be merged within SANS-CNN to provide even more robust SANS detection with additional choroidal vasculature data. Further research of SANS-CNN includes employing additional terrestrial neuro-ocular pathologies (e.g., papilledema in idiopathic intracranial hypertension) to further understand and delineate the similarities, differences, and spatial features of this unique microgravity phenomenon from terrestrial diseases. This may also help to further understand the etiology of SANS and its possible multi-factorial pathogenesis. Further prospective research with this machine learning model, including comparing astronaut terrestrial data to their SANS data, may further our understanding and ability to detect SANS.
Data availability
Due to the nature of the research, the supporting astronaut data is not available.
Code availability
Code available upon reasonable request.
References
Mader, T. H. et al. Optic disc edema, globe flattening, choroidal folds, and hyperopic shifts observed in astronauts after long-duration space flight. Ophthalmology 118, 2058–2069 (2011).
Lee, A. G., Mader, T. H., Gibson, C. R. & Tarver, W. Space flight-associated neuro-ocular syndrome. JAMA Ophthalmol. 135, 992–994 (2017).
Lee, A. G. et al. Spaceflight associated neuro-ocular syndrome (SANS) and the neuro-ophthalmologic effects of microgravity: a review and an update. NPJ Microgravity 6, 7 (2020).
Ong, J., Lee, A. G. & Moss, H. E. Head-down tilt bed rest studies as a terrestrial analog for spaceflight associated neuro-ocular syndrome. Front Neurol. 12, 648958, https://doi.org/10.3389/fneur.2021.648958 (2021).
Patel, N., Pass, A., Mason, S., Gibson, C. R. & Otto, C. Optical coherence tomography analysis of the optic nerve head and surrounding structures in long-duration international space station astronauts. JAMA Ophthalmol. 136, 193–200 (2018).
Laurie, S. S. et al. Optic disc Edema after 30 days of strict head-down tilt bed rest. Ophthalmology 126, 467–468 (2019).
Laurie, S. S. et al. Optic Disc Edema and choroidal engorgement in astronauts during spaceflight and individuals exposed to bed rest. JAMA Ophthalmol. 138, 165–172 (2020).
Zwart, S. R. et al. Association of genetics and B vitamin status with the magnitude of optic disc edema during 30-day strict head-down tilt bed rest. JAMA Ophthalmol. 137, 1195–1200 (2019).
Patel, Z. S. et al. Red risks for a journey to the red planet: The highest priority human health risks for a mission to Mars. NPJ Microgravity 6, 33 (2020).
Ong, J., Mader, T. H., Gibson, C. R., Mason, S. S. & Lee, A. G. Spaceflight associated neuro-ocular syndrome (SANS): an update on potential microgravity-based pathophysiology and mitigation development. Eye (Lond.) 37, 2409–2415 (2023).
Ong, J. & Lee, A. G. In Spaceflight Associated Neuro-Ocular Syndrome 1–7 (2022).
Lee, A. G., Mader, T. H., Gibson, C. R., Tarver, W. & Brunstetter, T. Lower body negative pressure as a potential countermeasure for spaceflight-associated neuro-ocular syndrome. JAMA Ophthalmol. 140, 652–653 (2022).
Hearon, C. M. Jr. et al. Effect of nightly lower body negative pressure on choroid engorgement in a model of spaceflight-associated neuro-ocular syndrome: a randomized crossover trial. JAMA Ophthalmol. 140, 59–65 (2022).
Scott, J. M. et al. Association of Exercise and swimming goggles with modulation of cerebro-ocular hemodynamics and pressures in a model of spaceflight-associated neuro-ocular syndrome. JAMA Ophthalmol. 137, 652–659 (2019).
Smith, S. M. & Zwart, S. R. Spaceflight-related ocular changes: the potential role of genetics, and the potential of B vitamins as a countermeasure. Curr. Opin. Clin. Nutr. Metab. Care 21, 481–488 (2018).
Lee, A. G. Optical coherence tomographic analysis of the optic nerve head and surrounding structures in space flight-associated neuro-ocular syndrome. JAMA Ophthalmol. 136, 200–201 (2018).
Ong, J. et al. Neuro-ophthalmic imaging and visual assessment technology for spaceflight associated neuro-ocular syndrome (SANS). Surv. Ophthalmol. 67, 1443–1466 (2022).
Ong, J. et al. Spaceflight associated neuro-ocular syndrome: proposed pathogenesis, terrestrial analogues, and emerging countermeasures. Br. J. Ophthalmol. 107, 895–900 (2023).
Seibert, M. A., Lim, D. S. S., Miller, M. J., Santiago-Materese, D. & Downs, M. T. Developing future deep-space telecommunication architectures: a historical look at the benefits of analog research on the development of solar system internetworking for future human spaceflight. Astrobiology 19, 462–477 (2019).
Ong, J. et al. Artificial intelligence frameworks to detect and investigate the pathophysiology of Spaceflight Associated Neuro-Ocular Syndrome (SANS). Brain Sci. 13, https://doi.org/10.3390/brainsci13081148 (2023).
Hossain, K. F., Kamran, S. A., Ong, J., Lee, A. G. & Tavakkoli, A. Revolutionizing Space Health (Swin-FSR): Advancing Super-Resolution of Fundus Images for SANS Visual Assessment Technology. Medical Image Computing and Computer Assisted Intervention – MICCAI 2023, 693–703, https://doi.org/10.1007/978-3-031-43990-2_65 (2023).
Chen, Y., Lu, W., Wang, J. & Qin, X. FedHealth 2: Weighted federated transfer learning via batch normalization for personalized healthcare. International Workshop on Federated and Transfer Learning for Data Sparsity and Confidentiality in Conjunction with IJCAI 2021 (FTL-IJCAI'21) (2021).
Kamran, S. A., Hossain, K. F., Tavakkoli, A. & Zuckerbrod, S. Attention2angiogan: Synthesizing fluorescein angiography from retinal fundus images using generative adversarial networks. 25th International Conference on Pattern Recognition (ICPR), 9122–9129, https://doi.org/10.1109/ICPR48806.2021.9412428 (2020).
Hossain, K. F. et al. ECG-Adv-GAN: Detecting ECG adversarial examples with conditional generative adversarial networks. 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), 50–56, https://doi.org/10.1109/icmla52953.2021.00016 (2021).
Chakraborty, S. Transfer learning based multi-fidelity physics informed deep neural network. J. Comput. Phys. 426, 109942 (2021).
Cai, C. et al. Transfer learning for drug discovery. J. Med. Chem. 63, 8683–8694 (2020).
Kamran, S. A. et al. New open-source software for subcellular segmentation and analysis of spatiotemporal fluorescence signals using deep learning. Iscience 25, 104277 (2022).
He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 770-778, https://doi.org/10.1109/CVPR.2016.90 (2016).
Krizhevsky, A., Sutskever, I. & Hinton, G. E. Imagenet classification with deep convolutional neural networks. Commun. ACM 60, 84–90 (2017).
Kamran, S. A. et al. RV-GAN: Segmenting retinal vascular structure in fundus photographs using a novel multi-scale generative adversarial network. Medical Image Computing and Computer Assisted Intervention – MICCAI 2021 (2021).
Tavakkoli, A., Kamran, S. A., Hossain, K. F. & Zuckerbrod, S. L. A novel deep learning conditional generative adversarial network for producing angiography images from retinal fundus photographs. Sci. Rep. 10, 1–15 (2020).
Howard, A. G. et al. Mobilenets: Efficient convolutional neural networks for mobile vision applications. Preprint at https://arxiv.org/abs/1704.04861 (2017).
Tan, M. & Le, Q. V. EfficientNet: Rethinking model scaling for convolutional neural networks. Proceedings of the 36th International Conference on Machine Learning (2019).
Kingma, D. P. & Ba, J. Adam: A method for stochastic optimization. 3rd International Conference for Learning Representations (ICLR) (2015).
Selvaraju, R. R. et al. Grad-cam: Visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE International Conference on computer vision, 618–626 (2017).
Chattopadhay, A., Sarkar, A., Howlader, P. & Balasubramanian, V. N. Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), 839–847 (2018).
Braithwaite, T. & Plant, G. T. Chronic optic disc swelling overlooked in a diabetic patient with a devastating outcome. BMJ Case Rep. 2010, https://doi.org/10.1136/bcr.06.2009.1975 (2010).
Olsen, T. W., Palejwala, N. V., Lee, L. B., Bergstrom, C. S. & Yeh, S. Chorioretinal folds: associated disorders and a related maculopathy. Am. J. Ophthalmol. 157, 1038–1047 (2014).
Masalkhi, M., Ong, J., Waisberg, E. & Lee, A. G. Chorioretinal folds in astronauts: risk of chorioretinal fold-related maculopathy and terrestrial staging of disease. Eye, https://doi.org/10.1038/s41433-023-02730-6 (2023).
Waisberg, E., Ong, J. & Lee, A. G. Anaemia, idiopathic intracranial hypertension (IIH) and spaceflight associated neuro-ocular syndrome (SANS). Eye, https://doi.org/10.1038/s41433-023-02834-z (2023).
Laurie, S. S. et al. Optic disc edema and chorioretinal folds develop during strict 6 degrees head-down tilt bed rest with or without artificial gravity. Physiol. Rep. 9, e14977, https://doi.org/10.14814/phy2.14977 (2021).
Waisberg, E. et al. Transfer learning as an AI-based solution to address limited datasets in space medicine. Life Sci. Space Res. 36, 36–38 (2023).
Waisberg, E. et al. Challenges of Artificial Intelligence in Space Medicine. Space: Science & Technology 2022, https://doi.org/10.34133/2022/9852872 (2022).
Paladugu, P. S. et al. Generative adversarial networks in medicine: important considerations for this emerging innovation in artificial intelligence. Ann. Biomed. Eng. 51, 2130–2142 (2023).
Zhang, Q., Wang, H., Lu, H., Won, D. & Yoon, S. W. In 2018 IEEE International Conference on Healthcare Informatics (ICHI) 199-207 (2018).
Ong, J. et al. Terrestrial health applications of visual assessment technology and machine learning in spaceflight associated neuro-ocular syndrome. NPJ Microgravity 8, 37 (2022).
Ong, J. et al. A multi-modal visual assessment system for monitoring Spaceflight Associated Neuro-Ocular Syndrome (SANS) during long duration spaceflight. J. Vis. 22, https://doi.org/10.1167/jov.22.3.6 (2022).
NASA. A non-intrusive ocular monitoring framework to model ocular structure and functional changes due to long-term spaceflight (80NSSC20K1831). NASA Life Sci. Data Arch. (2019).
Ong, J. et al. Head-mounted digital metamorphopsia suppression as a countermeasure for macular-related visual distortions for prolonged spaceflight missions and terrestrial health. Wear. Technol. 3, https://doi.org/10.1017/wtc.2022.21 (2022).
Sarker, P. et al. Extended reality quantification of pupil reactivity as a non-invasive assessment for the pathogenesis of spaceflight associated neuro-ocular syndrome: A technology validation study for astronaut health. Life Sci. Space Res. 38, 79–86 (2023).
Waisberg, E. et al. Dynamic visual acuity as a biometric for astronaut performance and safety. Life Sci. Space Res. 37, 3–6 (2023).
Acknowledgements
We would like to thank NASA and the NASA Human Research Program for their support of this research and providing the astronaut SANS OCT dataset through the through the NASA Lifetime Surveillance of Astronaut Health for validation of this machine learning model. This research has been presented as an abstract at the 2022 Optica Fall Vision Meeting. The work was funded by NASA Grant [80NSSC20K183]: A Non-intrusive Ocular Monitoring Framework to Model Ocular Structure and Functional Changes due to Long-term Spaceflight. Research reported in this publication was supported in part by the National Institute of General Medical Sciences of the National Institutes of Health under grant number P30 GM145646 and by the National Science Foundation under grant numbers OAC-2201599 and OIA-2148788 and by NASA under grant numbers 80NSSC23M0210 and 80NSSC20K1831.
Author information
Authors and Affiliations
Contributions
S.A.K., K.F.H., J.O., A.T. conceptualized and planned the experimental design. S.A.K., K.F.H., J.O. wrote the main manuscript text. S.A.K., K.F.H. conducted experiments. S.A.K., K.F.H., and J.O. revised the manuscript. N.Z., E.W., P.P., A.G.L., and A.T. critically reviewed the manuscript for feedback and preparation for submission.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing Interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kamran, S.A., Hossain, K.F., Ong, J. et al. SANS-CNN: An automated machine learning technique for spaceflight associated neuro-ocular syndrome with astronaut imaging data. npj Microgravity 10, 40 (2024). https://doi.org/10.1038/s41526-024-00364-w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41526-024-00364-w