Abstract
Phototransduction involves changes in concentration of ions and other solutes within photoreceptors and in subretinal space, which affect osmotic pressure and the associated water flow. Corresponding expansion and contraction of cellular layers can be imaged using optoretinography (ORG), based on phase-resolved optical coherence tomography (OCT). Until now, ORG could reliably detect only photoisomerization and phototransduction in photoreceptors, primarily in cones under bright stimuli. Here, by employing a phase-restoring subpixel motion correction algorithm, which enables imaging of the nanometer-scale tissue dynamics during minute-long recordings, and unsupervised learning of spatiotemporal patterns, we discover optical signatures of the other retinal structures’ response to visual stimuli. These include inner and outer segments of rod photoreceptors, retinal pigment epithelium, and subretinal space in general. The high sensitivity of our technique enables detection of the retinal responses to dim stimuli: down to 0.01% bleach level, corresponding to natural levels of scotopic illumination. We also demonstrate that with a single flash, the optoretinogram can map retinal responses across a 12° field of view, potentially replacing multifocal electroretinography. This technique expands the diagnostic capabilities and practical applicability of optoretinography, providing an alternative to electroretinography, while combining structural and functional retinal imaging in the same OCT machine.
Similar content being viewed by others
Introduction
Phototransduction involves changes in concentration of ions and other solutes within photoreceptors and in subretinal space (SRS), which affect osmotic pressure and the associated water flow. Ion homeostasis of SRS is regulated by retinal pigment epithelium (RPE), a monolayer of cells between the photoreceptor outer segment and the choroid, and a variety of ocular diseases, such as age-related macular degeneration, retinitis pigmentosa, and Best’s vitelliform macular dystrophy, are associated with dysfunction of epithelial transport of RPE cells1,2,3. In vivo assessment of RPE ion transport relies mostly on c-wave in electroretinography (ERG)4 or electrooculography5, which require contact electrodes and provide rather low spatial resolution.
Optoretinography (ORG) is an emerging imaging technology for non-invasive optical probing of retinal physiology in vivo. It usually utilizes phase-resolved optical coherence tomography (OCT) to detect mechanical deformations of retinal cells in response to visual stimuli, for example, the deformation of photoreceptors’ outer segments (OS)6,7,8,9,10,11, which is associated with photoisomerization and phototransduction. Unlike conventional OCT, where its axial motion sensitivity is constrained by the bandwidth-limited axial resolution (typically >1 μm), axial sensitivity of the phase-resolved OCT is defined by the signal-to-noise ratio, potentially enabling the detection of nanometer-scale tissue dynamics12. ORG was recently employed to classify chromatic cone subtypes in human subjects based on their responses to flashes of various colors13, and its clinical potential was demonstrated in assessing the progression of retinitis pigmentosa14. Furthermore, considering the similarity in photoreceptors’ structure and function across mammals, small animals such as rodents, serve as convenient and cost-effective models for studying the mechanisms of photoreceptor degeneration using ORG6,15. In genetically modified animals, for example Gnat1 and Gnat2 mutants and knockouts, ORG can provide insights into the various stages of the phototransduction cascade16,17. In addition, unlike human patients, animals are usually free of other ocular pathologies, and their imaging under anesthesia experiences fewer motion artifacts.
Despite these potential benefits of conducting ORG studies in rodents, the axial resolution of near-infrared OCT (NIR-OCT) systems, typically limited to several micrometers in tissue, hinders the clear differentiation of rodent photoreceptor OS terminals and RPE, particularly given the interdigitation of microvilli with the OS. In a typical structural image captured by NIR-OCT, the OS, and RPE are compounded in a speckled layer (Fig. 1a). Moreover, since the backscattered light is coherently convolved with the point spread function of the optical system, a single pixel in the OCT image can contain a mixture of phase signals from multiple cells. As a result, separating the ORG signals from the OS and RPE presents a significant challenge.
In this paper, we present a robust, unsupervised learning approach to finding the hidden spatiotemporal patterns in the phase signals measured from speckles, which reveals optical signatures of the SRS, photoreceptor inner and outer segments, and RPE responses to light. Notably, image registration using our phase-restoring subpixel motion correction algorithm enables ORG recordings down to nanometer-scale for tens of seconds, as opposed to the typical few seconds in previous studies, allowing for detection of much slower and more subtle phenomena in retinal responses to visual stimuli. We demonstrate these light-evoked responses under various scotopic and photopic conditions and map the OS and SRS dynamics across a wide field using a single flash.
Results
Outer retina dynamics in response to visual stimuli
All ORG imaging experiments were performed in vivo using a custom-built NIR-OCT with an axial resolution of 2.0 μm in tissue. Cross-sectional (Fig. 1a) and volumetric scans were acquired in time sequences from a 12° field of view (FOV), corresponding to 0.8 mm on wild-type rat retinas18. A despeckled image, generated by averaging the neighboring B-scans from a rectangular region, was used to validate the delineation of retinal layers (Fig. 1b). As illustrated in Fig. 1, several hyperreflective layers were observed in the outer retina, including the external limiting membrane (ELM), the inner segment/outer segment junction (IS/OS), Bruch’s membrane (BrM), and a thick speckling layer between the IS/OS and the BrM, which we call the mixed layer for convenience. As confirmed by the despeckled image, the mixed layer comprised photoreceptors’ OS and RPE cells. To reveal the light-evoked dynamics in the outer retina, first we computed the changes in the optical path length (OPL) (phase difference) between the BrM and three hyperreflective bands in the outer retina, including ELM, IS/OS, and the top band of the mixed layer (the red color in the bottom right of Fig. 1a). To achieve high phase stability/sensitivity in vivo, one needs to correct the bulk tissue motion. For this purpose, we applied our recently developed phase-restoring motion correction method to register complex-valued OCT images with subpixel precision19. Regarding the uncorrected out-of-plane motion, our experiments demonstrated that even for one-minute-long recordings, the out-of-plane slow drift was typically below 2 μm (see Supplementary Discussion 1 and Supplementary Fig. 1)—much smaller than the beam diameter (12.2 μm, theoretical 1/e2 width, potentially even larger due to ocular aberrations). The phase uncertainty resulting from such uncorrected out-of-plane slow drift remained marginal, which allowed reliable measurements of retinal dynamics throughout prolonged ORG recordings (see Supplementary Discussion 1 and Supplementary Fig. 2). Phase traces extracted from the prolonged ORG experiments (55 s in duration) were spatially averaged across pixels within each band. Note that in this study, the temporal phase change was extracted from a target layer with respect to a reference layer, which could be located either anterior or posterior to the target layer (see Supplementary Method 1). For consistency in interpretation of the increase/decrease of OPL as the expansion/contraction between the two layers, a negative sign was added when converting the phase change into the OPL change if the target layer was anterior to the reference layer.
As shown in Fig. 1c, distance between ELM and BrM increases after the stimulus (1 ms, wavelength 500 nm, 0.18% bleach level) at a rate of about 26 nm/s, reaching 220 nm around 20 s, and then slowly recovers afterwards. The IS/OS layer rapidly (~1 s) moves away from BrM, bounces back within 2 s and then continues to slowly move away from the BrM. The top red band of the mixed layer (presumably OS tips) rapidly (~1 s) moves toward the BrM by about 35 nm, then moves away from it over the next 20 s, followed by a slower recovery afterwards.
Conventionally, ORG monitors the movement of OS tips relative to the IS/OS layer. Due to the ambiguity of multiple tissue layers in NIR-OCT, we recorded the OPL changes from several bands in the mixed layer, indicated by various colors at the bottom right of Fig. 1a, relative to the IS/OS. As shown in Fig. 1d, the OPL changes in different bands contain a mixture of two distinct signatures: (a) rapid (~1 s) expansion followed by a slower (~10 s) decline, and (b) slow (~20 s) expansion, followed by even slower recovery. The superposition of distinct signals necessitates further signal decomposition and classification to determine the tissue origin associated with each individual signal.
Unsupervised learning of spatiotemporal patterns for signal classification
The limited resolution of NIR-OCT leads to the mixing of light-evoked responses from both OS and RPE, thereby complicating the interpretation of the results. To identify distinct signal patterns in an unbiased manner and group them into distinct types, we projected individual phase traces onto a spatiotemporal feature space and employed unsupervised learning using agglomerative hierarchical clustering based on Ward’s criterion20. We then trained a support vector machine (SVM) model on these type labels within the established feature space, enabling the extraction of each cluster’s decision boundary and the classification of new phase traces using the same decision boundaries (Methods & Supplementary Fig. 3).
To account for inter-subject differences, we combined phase traces from five rats for feature extraction and subsequent unsupervised clustering (Supplementary Table 2 and Supplementary Table 3). We conducted preprocessing and principal component analysis (PCA) on the phase traces to extract their temporal features (Methods & Supplementary Fig. 4). We then constructed a three-dimensional (3D) spatiotemporal feature space using the distance to BrM (depth) as the spatial feature and the top two principal components (PCs) as the temporal features (Fig. 2a). Outliers in low-density regions were removed using a distance-based algorithm (gray dots in Fig. 2a).
To differentiate between two distinct signatures (Fig. 1d), we employed an agglomerative hierarchical clustering algorithm based on Ward’s criterion20. By thresholding the dendrogram at the solid black line in Fig. 2b, the remaining phase traces (pink dots in Fig. 2a) were grouped into three clusters in the spatiotemporal feature space (Fig. 2c). A transition band (brown dots) facilitated better separation between the two distinct light-evoked tissue dynamics (Supplementary Fig. 5). Representative signals (Fig. 2d, red and blue dots in Fig. 2c) were obtained by averaging the individual phase traces within each cluster and converting them into OPL changes (ΔOPL). The first type of signal (red, Type-I) exhibited a rapid increase, peaking around 0.5 s, followed by a gradual decrease and a negative overshoot after 2.5 s (Fig. 2d). This Type-I signal matches the ORG signals from the photoreceptor outer segments in the literature, which were reported at shorter recording durations8,10. The second type of signal (blue, Type-II) is characterized by a much slower rise, and its peak is not reached within the 3.5-second time range plotted in Fig. 2d.
Subsequently, we trained an SVM using the labels obtained from unsupervised clustering in Fig. 2c to set boundaries for Type-I (red surface) and Type-II (blue surface) signals in the spatiotemporal feature space (Fig. 3a). Using the pre-trained SVM, we successfully extracted Type-I and Type-II signals from new datasets. Type-I signals were located more anteriorly than Type-II signals, exhibiting a distribution close to normal (red line in Fig. 3b). Type-II signals were localized anterior to BrM, fitting a Gaussian function (blue curve in Fig. 3b). The locations of the peaks differed significantly (*P < 0.05, n = 48). The locations of Type-I and Type-II signals corresponded to the OS and RPE in the intensity profile, respectively, as determined by the despeckled image (yellow line in Fig. 3b). This observation suggested that the Type-I and Type-II signals correspond to the dynamics of the OS and RPE relative to IS/OS, respectively. We interpolated and clustered the phase traces from the prolonged recording using the SVM. The Type-I signal peaked within one second, underwent a negative undershoot peaking at 10 s, and returned to baseline within 30 s (Fig. 3c).
We further investigated the dynamics of the SRS, the extracellular space surrounding photoreceptors, and spanning from the apical membrane of RPE to the ELM21. The dynamics between IS/OS and ELM can be calculated by averaging the signals extracted from the pixels in the ELM layer, using the IS/OS layer as a reference. By summing up the Type-II signal (distance from RPE to IS/OS) and the OPL change between IS/OS and ELM, we obtained the SRS dynamics. As illustrated by the gray trace in Fig. 3c, the SRS signal increased more slowly, peaking at about 250 nm around 20 s, followed by an even slower recovery. In our subsequent quantitative analyses, we focus on OS and SRS dynamics.
Signal dependence on stimulus parameters
We further quantified the signals in scotopic and photopic conditions by plotting the amplitude and latency of the peak of OS expansion and slope (expansion rate) from the SRS response (Fig. 4a). In scotopic conditions (Fig. 4b), the peak amplitude of the OS signal increased logarithmically with the stimulus strengths, from ∆OPL of about 10 (1.72) nm [mean (SD)] at a bleach level of 0.002% to about 53 (8.68) nm when the flash intensity increased 140-fold to 0.28%. The peak latencies remained within 440–500 ms at a bleach level <0.1%, but increased to 878 (140) ms and 1063 (108) ms in response to stimuli at 0.19% and 0.28% bleach level, respectively. In contrast, the slope of the SRS signal increased rapidly with the stimulus intensity up to 0.06% bleach level, and stabilized at the level of about 25 nm/s after that. In photopic conditions (Fig. 4c), the retina was pre-illuminated (500 nm) for 5 min and the flash was at 0.28% bleach level. Background illuminance below 6 × 102 photons/(μm2 s) did not reduce the OS response, but it decreased 5-fold at background of 6 × 104 photons/(μm2 s). The OS peak latency was not altered with an even stronger background - up to 6 × 103 photons/(μm2 s), but it dropped by about 30% at 6 × 104 photons/(μm2 s). Similarly, SRS expansion rate was affected by a background illuminance only above 6 × 102 photons/(μm2 s), which reduced the slope by about 26% to 18.1 nm/s.
En-face functional maps
Using repeated volumetric scans, we mapped the SRS and OS dynamics, analogous to multifocal ERG, but with a single flash. Figure 5a displays an OCT volume with a 12° FOV, with the structural contrast in gray and angiographic contrast highlighted in red. The spatiotemporal distribution of SRS and OS signals was obtained at a 0.10% bleach level. The temporal resolution was 8 Hz, and the total recording time was 5 s with 1 s baseline. Figure 5b shows the OS and SRS signals at specific time points. Figure 5c maps the spatial distribution of the phase traces, with each grid representing the average response from a 1.2° × 0.48° area. In general, the OS and SRS signal maps show high-fidelity detection over the entire FOV, and high signal variance was observed underneath two large blood vessels, outlined by dashed lines in Fig. 5b, c. A video illustrating the spatiotemporal evolution of the OS and SRS signals can be found in Supplementary Movie 1.
Signals from SRS and OS, recorded over a wide field in response to a single flash, offer a convenient approach to diagnostic mapping, in contrast to the slow and low-resolution multifocal ERG. This technique expands the practical applicability of optoretinography to studies of not only photoreceptors but also the RPE’s control of water dynamics in SRS in health and disease, thus providing an alternative to ERG.
Discussion
After disambiguation of various signals in the outer retina, dynamics of its cellular layers can be described independently. Time derivative of the tissue deformation, i.e., its expansion rate, is very informative since it may reveal the water influx rate. As shown in Fig. 6, expansion rate of OS reaches its maximum of about 250 nm/s within 0.1 s after the stimulus and drops back to zero during the next 1 s. RPE, on the other hand, reaches the maximum contraction rate of about −10 nm/s at 2 s after the stimulus, leading to the maximum contraction of approximately 20 nm at 3 s, and slowly recovering afterwards. Expansion of SRS begins about 1.5 s after the stimulus, reaching a rate of 15 nm/s in the following 10 s, after which the expansion rate gradually decreases. SRS expansion may originate from water transport across the RPE in response to a decrease in K+ and an increase in Na+ concentrations due to phototransduction in photoreceptors, as observed earlier using other methods22,23. The inner segment is compressed by about 10 nm during 1.5 s, after which it follows the dynamics of SRS, albeit at a lower amplitude (100 nm expansion at maximum).
It is interesting to relate the ORG signals to ERG previously recorded in the same species – pigmented rats, in response to a white flash24, shown in Fig. 6e. The a-wave in ERG – the photoreceptors’ response to light, begins right after the flash (within 10 ms), corresponding to the beginning of the OS expansion in ORG. Continuation of the a-wave is obscured by the b-wave, which starts a few tens of ms later and corresponds to the electric current generated by the second-order retinal neurons. C-wave in ERG takes over after about 0.5 s and reaches its maximum at around 1.5 s. It corresponds to a potassium current through RPE, induced by the photoreceptors’ response to light. The RPE contraction in ORG, which reaches its maximum rate between 1 and 2 s, overlaps with the peak of c-wave (Fig. 6d, e). This optical signature opens the window into physiology of the photoreceptor-RPE interactions and the interphotoreceptor matrix. Water transport is limited by the membrane permeability and hence its dynamics is much slower than that of electric current. Using a model of the OS elongation due to osmotic imbalance during phototransduction6, we estimated the membrane permeability coefficient based on the OS expansion rate. The water permeability coefficient of the rod OS membrane was found to be 5.9 × 10–3 cm s–1, not too far from 2.6 × 10–3 cm s–1 measured earlier in vitro – see Supplementary Discussion 2 for details.
Intriguingly, the hyperpolarized RPE layer is getting compressed in this process, unlike swelling of the hyperpolarized outer segments driven by osmotic pressure changes in response to released osmolytes during phototransduction.
The responses of rods and cones to light vary in many aspects, including sensitivity, time constant, and light adaptation25. These differences may be attributed to different expressions of isoforms involved in the phototransduction cascade and distinct structural organizations of plasma membranes in rods and cones26,27. Until now, ORG was observed mainly in cone photoreceptors, since detecting rod signals in human subjects is more challenging because rods are generally smaller and densely packed around cones28,29,30. In rats, however, 97% of photoreceptors are rods31, and hence our OS signals are likely dominated by the rod outer segments. Adaptive optics (AO) enables imaging of single rods in peripheral human retina, and one AO-OCT study reported that a flash bleaching 0.05% rhodopsin resulted in rod OS elongation of about 60 nm, while using the same flash bleaching 0.2% opsin did not result in a detectable cone OS elongation10. These results agree with our observations: 0.06% bleach caused OS elongation by 20-30 nm in rats.
Several studies investigated prolonged ORG signals using different methodologies. Lu et al. studied the changes in retinal layer thickness after strong visual stimuli (>23% rhodopsin bleach) by segmenting the layers from OCT structural images32. They examined micrometer-level responses during a 30-min dark adaptation process after the visual stimulus, with a temporal resolution of 2.1 s and a baseline signal fluctuation of hundreds of nanometers. Zhang et al. investigated the OCT intensity profile change in depths by averaging numerous A-scans. They observed a slow (peak latency of 10–100 s, depending on the bleach level) increase in distance between IS/OS and BrM in response to light, and attributed it to the elongation of the OS6. Notably, this slow signal resembles our SRS expansion, while the actual OS signal in our observations and previous reports10 rises and recovers much faster. A follow-up study by Pijewska et al. demonstrated that the phase-based ORG signals enabled higher detection sensitivity than the intensity-based processing methods33. Their results, obtained using strong stimuli (100% rhodopsin bleach), showed almost linear expansions between the ELM and BrM over 40 s after the visual stimuli were delivered.
Another interesting finding is the undershoot of the OS signal, with its tens of seconds-long recovery (Fig. 6a), which was not reported in previous studies due to shorter observation time. It may be related to water transport from the OS to SRS when the osmolytes (Gαt, Gβ1γ1) rebind to the cell membrane during deactivation of phototransduction. Restoration of the OS osmotic pressure during expansion of the SRS may be accompanied by its slight (20 nm) compression, which recovers later along with the other cellular structures in the outer retina.
Expansion of the SRS, closely related to phototransduction and water transport via RPE, is larger than that of the OS. Such signature may serve as a more sensitive measure of retinal physiology, providing additional diagnostic insights for various diseases involving the outer retina. ORG conveniently combines structural and functional retinal imaging in the same machine.
Methods
Ethical statement
All experiments were conducted in accordance with guidelines and approvals from Institutional Animal Care and Use Committee (IACUC), SingHealth (2020/SHS/1574).
System setup
All the ORG imaging experiments in this study were performed using a custom-built spectral-domain OCT (Fig. 7a). This NIR-OCT system operated with a broadband superluminescent diode (cBLMD-T-850-HP-I, λc = 840 nm, ∆λ = 146 nm, Superlum, Ireland), providing an axial resolution of 2.0 μm in tissue. A spectrometer interfaced with a line-scan camera (Cobra-800, Octoplus, E2V) acquired the spectral interference fringes at an A-scan speed of 250 kHz, corresponding to an image depth of 1.07 mm in air. For posterior segment imaging in rodent eyes, a lens-based afocal telescope conjugated the axis of a galvo scanner to the pupil. To reduce the incident beam size and increase the scan angle, the telescope’s magnification was 0.17 (scan lens: 80 mm focal length; ocular lens: 30 mm + 25 mm focal length). The theoretical diffraction-limited beam diameter was 12.2 µm (1/e2 width) with a standard rat eye model34. A function generator (PCIe-6363, National Instruments, USA) synchronized the camera acquisition (both A-scan and B-scan acquisitions), galvanometer scanning, and visual stimulation. The data acquisition was controlled by an interface software developed using NI LabVIEW (19.0.1, National Instruments, TX, USA).
For visual stimulation, a LED (MCWHLP1, Thorlabs, USA) was collimated by an aspheric condenser lens, and a narrow bandpass filter (500 ± 5 nm, #65-694, Edmund Optics, Singapore) was used to reshape the spectrum to optimize the sensitivity to rhodopsin35. The LED’s response time was ~300 μs, equivalent to 6% of the B-scan frame acquisition time, and LED’s response waveform to trigger is shown in Supplementary Fig. 6. A 43.4° Maxwellian illumination was projected to the posterior eye to cover an area of approximately 6.75 mm2. The power and duration of the light stimulus were converted to the bleach percentages of rhodopsin using a published method36,37, detailed in Supplementary Method 2.
To generate a despeckled retinal image, volumetric raster scans (500 A-scans × 125 B-scans) were collected within a 12° × 0.3° (0.8 mm × 0.02 mm) rectangular FOV. The camera speed was set to 100 kHz to enhance the signal-to-noise ratio. Neighboring B-scans were aligned and then averaged into a single frame38.
Animal experiment protocol
Wild-type, Brown Norway rats (Rat Resource and Research Centre, male = 33, female =15, age: 6–16 weeks) were used with details listed in Supplementary Table 2. Animals were sedated using a ketamine/xylazine combination to better maintain retinal functional responses compared to other commonly used anesthetics, such as isoflurane and urethane, while minimizing eye motion39,40. Vital signals, including heart and respiration rates, were monitored throughout the imaging sessions. Animals were placed in a prone position with their head restrained stereotaxically. Two mydriasis drops, 1% Tropicamide (Alcon, Geneva, Switzerland) and 2.5% Phenylephrine (Alcon, Geneva, Switzerland), were administrated onto the cornea before the imaging. The cornea was moisturized by a balanced salt solution frequently throughout the imaging sessions.
M-opsin and rhodopsin have similar sensitivity spectra (peaking at approximately 500 nm), while S-opsin’s sensitivity peaks at 350 nm, with minimal overlap with rhodopsin sensitivity spectrum. Since M-opsin co-expression ratio decreases from ventral to dorsal35, the scanning region was limited to the dorsal area to minimize the influence of M-opsin in cones.
Details of the acquisition and stimulation protocols for ORG imaging are listed in Supplementary Table 3. Three scanning protocols were used in this study. In the first protocol, we acquired 1000 A-scans per B-scan and 200 B-scans per second, with a total acquisition time of 5 s. In the second protocol, the B-scan time interval was increased to 40 ms, and 1375 B-scans were acquired over a period of 55 s. The third protocol consisted of 40 repeated volumes (25 B-scans per volume) recorded at 8 volumetric scans per second for a period of 5 s. Flash intensity, dark/light adaptation, and inter-flash time interval varied in different experiments.
Automated extraction of the outer retina dynamics from speckle patterns
MATLAB (2020a, 2021a, and 2022b, MathWorks, MA, USA) was used for OCT image processing, registration, and all subsequent phase signal analysis. Python (v3.9.12) was used for plotting figures.
Preprocessing of OCT images and phase traces
The raw interference fringe first underwent standard OCT post-processing steps, including spectral calibration, k-linearity, dispersion compensation, and discrete Fourier transform (DFT), which resulted in complex-valued OCT images.
Phase-resolved OCT is very susceptible to bulk tissue motion, which results in apparent image distortions in the time-elapsed intensity M-scan and degrade the phase stability (see Fig. 7c). To correct the motion-induced phase error, we estimated subpixel-level translational displacements between repeated B-scans using the single-step DFT algorithm (see Fig. 7d)41. We registered the complex-valued OCT images using our recently developed phase-restoring subpixel motion correction algorithm19, where the lateral and axial displacements were corrected by multiplying the corresponding exponential terms in the spatial frequency domain and the spectrum domain, respectively. After image registration (see Fig. 7e), we achieved excellent motion stability at photoreceptor layers (gray arrows), while periodical oscillations from the vascular pulsation (pink arrows) could be observed in the choroid.
Light-evoked dynamics of the outer retina was then extracted by computing the temporal phase difference between the pixel pairs from the outer retinal bands. Several hyperreflective bands, including ELM, IS/OS, BrM, and a thick speckling layer that contains photoreceptors’ OS and RPE cells, were automatically segmented using graph theory and dynamic programming (Supplementary Fig. 3a)42. In self-referencing measurements, for each pixel in a target layer, we computed its temporal phase change with respect to a reference layer, detailed in Supplementary Method 1.
We also conducted temporal filtering to remove unwanted signal frequencies. The extracted signals were first processed by bandstop filters to minimize residual artifacts induced by heartbeat and breathing. A low-pass filter with a cut-off frequency of 10 Hz was subsequently used to filter out high-frequency oscillations. After filtering, both ends of the phase traces exhibited high variance due to edge effects, so they were removed from the data analysis afterwards (Supplementary Fig. 3c). For individual phase traces, our filters effectively captured the signal profiles while suppressing both periodic oscillations and high-frequency noise (Supplementary Fig. 3b). Meanwhile, it should be noted that the low-pass filter was implemented exclusively for the signal classification. After identifying pixels corresponding to the Type-I and Type-II signals, the light-evoked responses reported in this article were not processed by any low-pass filter to prevent potential signal distortion.
Construction of spatiotemporal feature space
We used the PCA to compress high-dimensional phase traces into a lower-dimensional feature space to facilitate the analysis (Supplementary Fig. 4c). Traces with a pre-stimulus SD larger than 60 mrad were excluded, mostly from pixels with low SNR or underneath blood vessels. Then each trace was normalized by subtracting its mean value and dividing by its SD (Supplementary Fig. 4). To account for inter-subject differences, we combined normalized traces extracted from five rats to calculate PC coefficients. The top two PCs capturing a total variance of 73.3% were selected as temporal features and its axial distance to the BrM was selected as the spatial feature. We normalized both the top two PCs and the depth information to construct a three-dimensional spatiotemporal feature space.
To avoid undesirable elongated clusters in subsequent unsupervised clustering analysis, we used a distance-based outlier detection method to preclude outliers distributed in low-density regions in the spatiotemporal feature space43,44. For each data point in the feature space, we calculated the minimum radius of a sphere that is centered at that point and can cover 2% of the remaining data points. This radius reflected the local distribution density around each data point in the feature space. Then, a given point would be labeled as an outlier if its corresponding radius was larger than \({Q}_{3}+\left({Q}_{3}-{Q}_{1}\right)/5\), where Q1 and Q3 are the first and third quartiles of the calculated radii.
Unsupervised clustering with a hierarchical clustering algorithm
We used an agglomerative hierarchical clustering algorithm under the Ward criterion to group individual points in the established spatiotemporal feature space20. The Euclidean distance of each pair of points was computed and similar subclusters were then iteratively merged into larger clusters. Under the guidance of the Ward method or minimum variance method, each merger guaranteed a minimum increase of total within-cluster variance.
Training a support vector model in the established feature space
We trained a SVM in the spatiotemporal feature space with the previously obtained labels, including outlier, intermediate phase trace, Type-I signal, and Type-II signal, to obtain their decision boundaries. The input features were standardized before training, the Gaussian kernel was selected, and automatic hyperparameters optimization was turned on. Evaluated with the 10-fold cross-validation strategy, the trained SVM achieved a classification accuracy of 99.2%. The SVM is useful for processing a new dataset using the same criterion. Phase traces extracted from the new dataset will be preprocessed and projected onto the same feature space. Then Type-I and Type-II signals can be automatically extracted based on the classification results from the SVM.
Processing of datasets with lower temporal resolution
Prolonged recording (protocol 2) and volumetric scans (protocol 3) had an 8-fold and 25-fold lower temporal sampling rate than repeated B-scans (protocol 1), respectively. To cluster these phase traces in the same spatiotemporal feature space, they were first linearly interpolated into 5 ms time intervals and smoothed using a Gaussian filter. Other procedures were similar to those used for processing the repeated B-scans datasets.
Statistics and reproducibility
Sample size calculation was not conducted for the pilot animal experiments. Comparison of depth between Type I and II signals was analyzed using paired, two-tailed Student’s t test, after confirming normality with the Shapiro-Wilk test. p < 0.05 was considered statistically significant. All statistical analyses were executed using MATLAB (R2020b, MathWorks).
Animals were randomly assigned to different study groups. Data with excessive motion artifacts (mainly due to heavier breathing effects of certain animals during anesthesia) were excluded from further analysis, because they introduced severe out-of-plane movement, which our image registration algorithm could not correct.
Reporting summary
Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.
Data availability
All data needed to evaluate the conclusions in the paper are present in the paper and the Supplementary Information. Two example datasets, one corresponding to a 5-second recording with an acquisition rate of 200 B-scans/second and the other corresponding to a prolonged recording with an acquisition rate of 25 B-scans/second, can be downloaded from https://github.com/NTU-Ling-lab/ORG-Classification. The raw experimental data are too large to be publicly shared, yet they are available for research purposes from corresponding authors upon request. Requests will be fulfilled within 2 months.Source data are provided with this paper.
Code availability
The algorithm demo for our prolonged and multilayered optoretinography (ORG) method can be found at https://github.com/NTU-Ling-lab/ORG-Classification45.
References
Strauss, O. The retinal pigment epithelium in visual function. Physiol. Rev. 85, 845–881 (2005).
Sparrow, J. R., Hicks, D. & Hamel, C. P. The retinal pigment epithelium in health and disease. Curr. Mol. Med. 10, 802–823 (2010).
Lakkaraju, A. et al. The cell biology of the retinal pigment epithelium. Prog. Retinal Eye Res. 78, 100846 (2020).
Oakley n, B. & Green, D. G. Correlation of light-induced changes in retinal extracellular potassium concentration with c-wave of the electroretinogram. J. Neurophysiol. 39, 1117–1133 (1976).
Joseph, D. P. & Miller, S. S. Apical and basal membrane ion transport mechanisms in bovine retinal pigment epithelium. J. Physiol. 435, 439–463 (1991).
Zhang, P. et al. In vivo optophysiology reveals that G-protein activation triggers osmotic swelling and increased light scattering of rod photoreceptors. Proc. Natl Acad. Sci. 114, E2937–E2946 (2017).
Pandiyan, V. P. et al. The optoretinogram reveals the primary steps of phototransduction in the living human eye. Sci. Adv. 6, eabc1124 (2020).
Hillmann, D. et al. In vivo optical imaging of physiological responses to photostimulation in human photoreceptors. Proc. Natl Acad. Sci. 113, 13138–13143 (2016).
Pandiyan, V. P., Nguyen, P. T., Pugh, E. N. Jr & Sabesan, R. Human cone elongation responses can be explained by photoactivated cone opsin and membrane swelling and osmotic response to phosphate produced by RGS9-catalyzed GTPase. Proc. Natl Acad. Sci. USA 119, e2202485119 (2022).
Azimipour, M. et al. Optoretinogram: optical measurement of human cone and rod photoreceptor responses to light. Opt. Lett. 45, 4658–4661 (2020).
Tomczewski, S. et al. Light-adapted flicker optoretinograms captured with a spatio-temporal optical coherence-tomography (STOC-T) system. Biomed. Opt. Express 13, 2186–2201 (2022).
Goodman J. W. Statistical optics, (John Wiley & Sons, 2015).
Zhang, F., Kurokawa, K., Lassoued, A., Crowell, J. A. & Miller, D. T. Cone photoreceptor classification in the living human eye from photostimulation-induced phase dynamics. Proc. Natl Acad. Sci. 116, 7951–7956 (2019).
Lassoued, A. et al. Cone photoreceptor dysfunction in retinitis pigmentosa revealed by optoretinography. Proc. Natl Acad. Sci. USA 118, e2107444118 (2021).
Kim, T. H., Wang, B., Lu, Y., Son, T. & Yao, X. Functional optical coherence tomography enables in vivo optoretinography of photoreceptor dysfunction due to retinal degeneration. Biomed. Opt. Express 11, 5306–5320 (2020).
Chang, B. et al. Cone photoreceptor function loss-3, a novel mouse model of achromatopsia due to a mutation in Gnat2. Investig. Ophthalmol. Vis. Sci. 47, 5017–5021 (2006).
Calvert, P. et al. Phototransduction in transgenic mice after targeted deletion of the rod transducin α-subunit. Proc. Natl Acad. Sci. 97, 13913–13918 (2000).
Wang, B.-Y. et al. Electronic photoreceptors enable prosthetic visual acuity matching the natural resolution in rats. Nat. Commun. 13, 6627 (2022).
Li, H. et al. Shot-noise limited phase-sensitive imaging of moving samples by phase-restoring subpixel motion correction in Fourier-domain optical coherence tomography. Preprint at bioRxiv 2022.06.15.496241, https://doi.org/10.1101/2022.06.15.496241 (2022).
Ward, J. H. Jr Hierarchical grouping to optimize an objective function. J. Am. Stat. Assoc. 58, 236–244 (1963).
Lindell, M. et al. Volumetric reconstruction of a human retinal pigment epithelial cell reveals specialized membranes and polarized distribution of organelles. Investig. Ophthalmol. Vis. Sci. 64, 35–35 (2023).
Huang, B. & Karwoski, C. Light-evoked expansion of subretinal space volume in the retina of the frog. J. Neurosci. 12, 4243–4252 (1992).
Dmitriev, A. V., Govardovskii, V. I., Schwahn, H. N. & Steinberg, R. H. Light-induced changes of extracellular ions and volume in the isolated chick retina-pigment epithelium preparation. Vis. Neurosci. 16, 1157–1167 (1999).
Tan, B., Mason, E., MacLellan, B. & Bizheva, K. K. Correlation of visually evoked functional and blood flow changes in the rat retina measured with a combined OCT+ ERG system. Investig. Ophthalmol. Vis. Sci. 58, 1673 (2017).
Korenbrot, J. I. Speed, sensitivity, and stability of the light response in rod and cone photoreceptors: Facts and models. Prog. Retin Eye Res. 31, 442–466 (2012).
Ingram, N. T., Sampath, A. P. & Fain, G. L. Why are rods more sensitive than cones? J. Physiol. 594, 5415–5426 (2016).
Mustafi, D., Engel, A. H. & Palczewski, K. Structure of cone photoreceptors. Prog. Retin Eye Res. 28, 289–302 (2009).
Dubra, A. et al. Noninvasive imaging of the human rod photoreceptor mosaic using a confocal adaptive optics scanning ophthalmoscope. Biomed. Opt. Express 2, 1864–1876 (2011).
Doble, N. et al. In vivo imaging of the human rod photoreceptor mosaic. Opt. Lett. 36, 31–33 (2011).
Liu, Z., Zhang, F., Zucca, K., Agrawal, A. & Hammer, D. X. Ultrahigh-speed multimodal adaptive optics system for microscopic structural and functional imaging of the human retina. Biomed. Opt. Express 13, 5860–5878 (2022).
Fu, Y. & Yau, K.-W. Phototransduction in mouse rods and cones. Pflügers Arch. - Eur. J. Physiol. 454, 805–819 (2007).
Lu, C. D. et al. Photoreceptor layer thickness changes during dark adaptation observed with ultrahigh-resolution optical coherence tomography. Investig. Ophthalmol. Vis. Sci. 58, 4632–4643 (2017).
Pijewska, E. et al. Extraction of phase-based optoretinograms (ORG) from serial B-scans acquired over tens of seconds by mouse retinal raster scanning OCT system. Biomed. Opt. Express 12, 7849–7871 (2021).
Campbell, M. & Hughes, A. An analytic, gradient index schematic lens and eye for the rat which predicts aberrations for finite pupils. Vis. Res. 21, 1129–1148 (1981).
Wang, Y. V., Weick, M. & Demb, J. B. Spectral and temporal sensitivity of cone-mediated responses in mouse retinal ganglion cells. J. Neurosci. 31, 7670–7681 (2011).
Perlman, I. Kinetics of bleaching and regeneration of rhodopsin in abnormal (RCS) and normal albino rats in vivo. J. Physiol. 278, 141–159 (1978).
Zhang, P., Goswami, M., Zawadzki, R. J. & Pugh, E. N. Jr The photosensitivity of rhodopsin bleaching and light-induced increases of fundus reflectance in mice measured in vivo with scanning laser ophthalmoscopy. Investig. Ophthalmol. Vis. Sci. 57, 3650–3664 (2016).
Kho, A. M., Zhang, T., Zhu, J., Merkle, C. W. & Srinivasan, V. J. Incoherent excess noise spectrally encodes broadband light sources. Light Sci. Appl. 9, 172 (2020).
Choh, V. et al. Isoflurane and ketamine:xylazine differentially affect intraocular pressure-associated scotopic threshold responses in Sprague-Dawley rats. Doc. Ophthalmol. 135, 121–132 (2017).
Nair, G. et al. Effects of common anesthetics on eye movement and electroretinogram. Doc. Ophthalmol. 122, 163–176 (2011).
Guizar-Sicairos, M., Thurman, S. T. & Fienup, J. R. Efficient subpixel image registration algorithms. Opt. Lett. 33, 156–158 (2008).
LaRocca, F. et al. Robust automatic segmentation of corneal layer boundaries in SDOCT images using graph theory and dynamic programming. Biomed. Opt. Express 2, 1524–1538 (2011).
Ros, F. & Guillaume, S. A hierarchical clustering algorithm and an improvement of the single linkage criterion to deal with noise. Expert Syst. Appl. 128, 96–108 (2019).
Knorr, E. M. & Ng R. T. Algorithms for mining distance-based outliers in large datasets. Proceedings of the 24rd International Conference on Very Large Data Bases, 392–403 (Morgan Kaufmann Publishers Inc., 1998).
Li H., Ling T. Optoretinography classification method, NTU-Ling-lab, GitHub.com, https://doi.org/10.5281/zenodo.11190406, (2024).
Acknowledgements
We thank Drs. R. Sabesan, A. Roorda, E. Pugh, and M. Ismail for fruitful discussions. This work was funded by grants from National Research Foundation Singapore (NRF-NRFF14-2022-0005, T.L.; NRF2019-THE002-0006, L.S.; NRF-CRP24-2020-0001, L.S.), the Startup Grant from Nanyang Technological University (T.L.), National Medical Research Council (NMRC/CG/C010A/2017, T.L.; CG/C010A/2017_SERI, L.S.; OFIRG/0048/2017, L.S.; OFLCG/004c/2018, L.S.; TA/MOH-000249-00/2018, L.S.; MOH-OFIRG20nov-0014, L.S.; NMRC/CG/M010/2017/Pre-Clinical, V.A.B.), the Ministry of Education, Singapore under its AcRF Tier 1 Grant (RS19/20, T.L.; RG28/21, T.L.), A*STAR (A20H4b0141, L.S.), the Singapore Eye Research Institute & Nanyang Technological University (SERI-NTU Advanced Ocular Engineering (STANCE) Program, L.S.), the Duke-NUS Medical School (Duke-NUS-KP(Coll)/2018/0009 A, L.S.), the SERI-Lee Foundation (LF1019-1, L.S.), NIH grant (U01 EY025501, D.P.) and AFOSR grant (FA9550-20-1-0186, D.P.).
Author information
Authors and Affiliations
Contributions
B.T., H.L., D.P., L.S., and T.L. designed the study. B.T. and L.S. built the NIR-OCT setup. B.T., H.L., D.N., and L.H. conducted the experiments. V.A.B. supported the animal preparation. B.T., H.L., R.M., L.H., and T.L. analyzed the data. B.T., H.L., Y.Z., D.P., L.S., and T.L. wrote the draft. All work was supervised by T.L., D.P., and L.S. T.L. and L.S. obtained funding. All authors contributed to the final manuscript.
Corresponding authors
Ethics declarations
Competing interests
B.T., H.L., L.S., and T.L. are inventors on a PCT patent application (PCT/SG2024/050050) related to unsupervised signal classification and processing for optoretinography. The other authors declare no competing interests.
Peer review
Peer review information
Nature Communications thanks Guangying Ma and Omar Mahroo for their contribution to the peer review of this work. A peer review file is available.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Source data
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Tan, B., Li, H., Zhuo, Y. et al. Light-evoked deformations in rod photoreceptors, pigment epithelium and subretinal space revealed by prolonged and multilayered optoretinography. Nat Commun 15, 5156 (2024). https://doi.org/10.1038/s41467-024-49014-5
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-024-49014-5
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.