Hostname: page-component-cd9895bd7-7cvxr Total loading time: 0 Render date: 2024-12-29T10:13:25.988Z Has data issue: false hasContentIssue false

Visual Snow Syndrome: Use of Text-To-Image Artificial Intelligence Models to Improve the Patient Perspective

Published online by Cambridge University Press:  10 November 2022

Michael Balas
Affiliation:
Temerty Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada
Jonathan A. Micieli*
Affiliation:
Department of Ophthalmology and Vision Sciences, University of Toronto, Toronto, Ontario, Canada Division of Neurology, Department of Medicine, St. Michael’s Hospital, University of Toronto, Toronto, Ontario, Canada Kensington Vision and Research Centre, Toronto, Ontario, Canada
*
Corresponding author: Jonathan A. Micieli, Kensington Vision and Research Centre, 340 College Street, Suite 501, Toronto, Ontario M5T 3A9, Canada. Email: jmicieli@kensingtonhealth.org
Rights & Permissions [Opens in a new window]

Abstract

Type
Letter to the Editor: New Observation
Copyright
© The Author(s), 2022. Published by Cambridge University Press on behalf of Canadian Neurological Sciences Federation

Visual snow syndrome (VSS) is a positive visual disturbance that has been described as the constant flickering of innumerable tiny dots throughout the entire visual field. In other words, it is like seeing the world through the static noise of an untuned television. Reference Puledda, Schankin and Goadsby1 VSS has only been identified in the literature recently, with the first case report written by Liu et al. in 1995. Reference Liu, Schatz, Galetta, Volpe, Skobieranda and Kosmorsky2 The current criteria to make a diagnosis of VSS include the presence of tiny dots in the entire visual field lasting more than 3 months, in addition to the presence of at least two of the following visual symptoms: palinopsia, entoptic phenomena, photophobia, and nyctalopia. Reference Puledda, Schankin and Goadsby1 Moreover, the symptoms should not be consistent with typical migraine visual aura or better explained by another disorder, such as hallucinogen persisting perception disorder. This syndrome typically presents in young adulthood, with the most common comorbid conditions being migraine and tinnitus. Although VSS is an increasingly recognized condition, it may be hard for the clinician to truly appreciate the patient perspective and to have a pictorial representation of the symptoms. Artificial intelligence (AI) allows for a novel approach to create illustrations based on a combination of words. We made use of various text-to-image algorithms and applied them to a patient with VSS.

A 17-year-old man was referred from his optometrist due to complaints of “seeing everything through grainy static interference in both eyes”, which he first noticed two years prior to presentation. This static vision was constant and more obvious in light-coloured environments, such as looking at a white wall. He also endorsed photosensitivity in addition to a floater noticed in the left eye that was particularly visible in bright environments. He described certain instances where he had poor vision at night, and he reported no palinopsia. His past medical history includes a myringotomy at age 4 for inner ear blockages that have resolved since age 9, as well as sinus headaches that occur with environmental pressure changes. Both of his parents experienced migraines. He has no allergies and reports no alcohol use, smoking, or recreational drug use including hallucinogens.

The examination revealed normal afferent visual function including normal visual acuity (20/20 OU) and full confrontation visual fields. Pupillary examination revealed normal sized pupils without a relative afferent pupillary defect, and dilated examination of the fundus demonstrated normal appearing optic nerves and maculae. Optical coherence tomography of the retinal nerve fiber layer and macula in both eyes were also normal. Humphrey 24-2 SITA-Fast visual fields were normal.

We performed an additional phone interview with the patient to obtain more information and descriptive terms of the visual phenomena he was experiencing, in his own words. We then took his self-reported descriptions and translated them into images using various publicly available text-to-image AI models, including “DALL · E 2” (www.openai.com/dall-e-2), “Stable Diffusion” (www.stability.ai), and “Midjourney” (www.midjourney.com/home). These models are all freely accessible (up to a certain number of generated images) via their own respective websites and easy to use, with no prior computer science experience required. After examining all of the generated images from each model, we selected the top two that we believed were most appropriate and consistent with the patient’s story as well as previous artistic renditions of VSS. These images are shown in Figures 1 and 2. They were generated within seconds using the “Midjourney” model from a short description of a complex visual syndrome in a realistic environment. While this technology continues to improve and evolve, our results already highlight the promise for AI to help us better understand our patients and see the world through their eyes.

Figure 1: Generated using Midjourney with the caption: “POV lying in bed at night seeing through grainy static interference.” This image portrays the patient’s first time experiencing these symptoms.

Figure 2: Generated using Midjourney with the caption “Seeing a snowy landscape through static grainy interference.” This image portrays the circumstances which elicit the patient’s visual symptoms the most: snow, clouds, and other environments with bright white backgrounds.

VSS is a relatively new condition that does not have any objective examination findings. Reference Metzler and Robertson3 A diagnosis is made based on the patient’s description of their experience. This may be difficult for family members, friends, and medical staff to understand, since patients may use variable language and a pictorial representation of their experience is difficult to obtain. In this paper, we made use of various text-to-image AI models and applied them to VSS to produce images that can be used in educational settings.

“DALL · E 2,” “Stable Diffusion,” and “Midjourney,” among other text-to-image models, all comprise different variations of generative models and function in similar ways, seamlessly converting user-specified text into evocative images. Some of the current state-of-the-art model architectures include Transformer Language Models which derive meaning from the textual input, typically used in conjunction with generative adversarial networks (GANs), variational autoencoders (VAEs), or diffusion probabilistic models (DMs), which synthesize the image based on the transformer’s text embeddings. Reference Rombach, Blattmann, Lorenz, Esser and Ommer4Reference Ramesh, Pavlov and Goh6 These text-to-image training methods are also often described as Zero-Shot Learning, which refers to the process by which these AI models are able to understand and generate data they have never seen before.

Given their novelty, text-to-image models and similar technologies have yet to be used in the medical literature or integrated into clinical practice. In this report, we describe one possible application of this technology, which is to empower patients with the ability to translate their concerns into rendered visual images. This can help patients clarify their ailments as well as foster empathy by helping them feel understood. Another possible use case of this technology involves the opposite, whereby providers can better communicate visual disease or treatment outcomes to the patient, such as in plastic reconstructive surgery to set expectations or in conditions with physical manifestations to promote medication adherence and lifestyle modifications. These text-to-image models may also be used to generate scientific or educational materials that can be incorporated into textbooks and presentations without requiring licensing agreements and illustrators.

Ultimately, these AI models introduce a new dimension of communication into patient care and medicine, one that will continue to grow and evolve in tandem with the sophistication of these algorithms. We encourage the early adoption and use of these technologies to ensure that medicine keeps pace and helps inform their development.

Funding

This work was not funded.

Conflict of Interest

The authors have no conflicts of interest to declare.

Disclosures

The authors have no relevant disclosures.

Statement of authorship

All authors attest that they meet the current International Committee of Medical Journal Editors criteria for authorship.

References

Puledda, F, Schankin, C, Goadsby, PJ. Visual snow syndrome: a clinical and phenotypical description of 1,100 cases. Neurology. 2020;94:e564e574.CrossRefGoogle ScholarPubMed
Liu, G, Schatz, N, Galetta, S, Volpe, N, Skobieranda, F, Kosmorsky, G. Persistent positive visual phenomena in migraine. Neurology. 1995;45:6648.CrossRefGoogle ScholarPubMed
Metzler, AI, Robertson, CE. Visual snow syndrome: proposed criteria, clinical implications, and pathophysiology. Curr Neurol Neurosci Rep. 2018;18:19.CrossRefGoogle ScholarPubMed
Rombach, R, Blattmann, A, Lorenz, D, Esser, P, Ommer, B. High-resolution image synthesis with latent diffusion models. In: Paper presented at: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022.CrossRefGoogle Scholar
Saharia, C, Chan, W, Saxena, S, et al. Photorealistic text-to-image diffusion models with deep language understanding, arXiv preprint arXiv: 220511487. 2022.Google Scholar
Ramesh, A, Pavlov, M, Goh, G, et al. Zero-shot text-to-image generation. In: Paper presented at: International Conference on Machine Learning 2021.Google Scholar
Figure 0

Figure 1: Generated using Midjourney with the caption: “POV lying in bed at night seeing through grainy static interference.” This image portrays the patient’s first time experiencing these symptoms.

Figure 1

Figure 2: Generated using Midjourney with the caption “Seeing a snowy landscape through static grainy interference.” This image portrays the circumstances which elicit the patient’s visual symptoms the most: snow, clouds, and other environments with bright white backgrounds.