Summary

Quantification of Visual Feature Selectivity of the Optokinetic Reflex in Mice

Published: June 23, 2023
doi:

Summary

Here, we describe a standard protocol for quantifying the optokinetic reflex. It combines virtual drum stimulation and video-oculography, and thus allows precise evaluation of the feature selectivity of the behavior and its adaptive plasticity.

Abstract

The optokinetic reflex (OKR) is an essential innate eye movement that is triggered by the global motion of the visual environment and serves to stabilize retinal images. Due to its importance and robustness, the OKR has been used to study visual-motor learning and to evaluate the visual functions of mice with different genetic backgrounds, ages, and drug treatments. Here, we introduce a procedure for evaluating OKR responses of head-fixed mice with high accuracy. Head fixation can rule out the contribution of vestibular stimulation on eye movements, making it possible to measure eye movements triggered only by visual motion. The OKR is elicited by a virtual drum system, in which a vertical grating presented on three computer monitors drifts horizontally in an oscillatory manner or unidirectionally at a constant velocity. With this virtual reality system, we can systematically change visual parameters like spatial frequency, temporal/oscillation frequency, contrast, luminance, and the direction of gratings, and quantify tuning curves of visual feature selectivity. High-speed infrared video-oculography ensures accurate measurement of the trajectory of eye movements. The eyes of individual mice are calibrated to provide opportunities to compare the OKRs between animals of different ages, genders, and genetic backgrounds. The quantitative power of this technique allows it to detect changes in the OKR when this behavior plastically adapts due to aging, sensory experience, or motor learning; thus, it makes this technique a valuable addition to the repertoire of tools used to investigate the plasticity of ocular behaviors.

Introduction

In response to visual stimuli in the environment, our eyes move to shift our gaze, stabilize retinal images, track moving targets, or align the foveae of two eyes with targets located at different distances from the observer, which are vital to proper vision1,2. Oculomotor behaviors have been widely used as attractive models of sensorimotor integration to understand the neural circuits in health and disease, at least partly because of the simplicity of the oculomotor system3. Controlled by three pairs of extraocular muscles, the eye rotates in the socket primarily around three corresponding axes: elevation and depression along the transverse axis, adduction and abduction along the vertical axis, and intorsion and extorsion along the anteroposterior axis1,2. Such a simple system allows researchers to evaluate the oculomotor behaviors of mice easily and accurately in a lab environment.

One prime oculomotor behavior is the optokinetic reflex (OKR). This involuntary eye movement is triggered by slow drifts or slips of images on the retina and serves to stabilize retinal images as an animal's head or its surroundings move2,4. The OKR, as a behavioral paradigm, is interesting to researchers for several reasons. First, it can be stimulated reliably and quantified accurately5,6. Second, the procedures of quantifying this behavior are relatively simple and standardized and can be applied to evaluate the visual functions of a large cohort of animals7. Third, this innate behavior is highly plastic5,8,9. Its amplitude can be potentiated when repetitive retinal slips occur for a long time5,8,9, or when its working partner vestibular ocular reflex (VOR), another mechanism of stabilizing retinal images triggered by vestibular input2, is impaired5. These experimental paradigms of OKR potentiation empower researchers to unveil the circuit basis underlying oculomotor learning.

Two non-invasive methods have primarily been used to evaluate the OKR in previous studies: (1) video-oculography combined with a physical drum7,10,11,12,13 or (2) arbitrary determination of head turns combined with a virtual drum6,14,15,16. Although their applications have made fruitful discoveries in understanding the molecular and circuit mechanisms of oculomotor plasticity, these two methods each have some drawbacks which limit their powers in quantitatively examining the properties of the OKR. First, physical drums, with printed patterns of black and white stripes or dots, do not allow easy and quick changes of visual patterns, which largely restricts the measurement of the dependence of the OKR on certain visual features, such as spatial frequency, direction, and contrast of moving gratings8,17. Instead, tests of the selectivity of the OKR to these visual features can benefit from computerized visual stimulation, in which visual features can be conveniently modified from trial to trial. In this way, researchers can systematically examine the OKR behavior in the multi-dimensional visual parameter space. Moreover, the second method of the OKR assay reports only the thresholds of visual parameters that trigger discernible OKRs, but not the amplitudes of eye or head movements6, 14,15,16. The lack of quantitative power thus prevents analyzing the shape of tuning curves and the preferred visual features, or detecting subtle differences between individual mice in normal and pathological conditions. To overcome the above limitations, video-oculography and computerized virtual visual stimulation had been combined to assay the OKR behavior in recent studies5,17,18,19,20. However, these previously published studies did not provide enough technical details or step-by-step instructions, and consequently it is still challenging for researchers to establish such an OKR test for their own research.

Here, we present a protocol to precisely quantify the visual feature selectivity of OKR behavior under photopic or scotopic conditions with the combination of video-oculography and computerized virtual visual stimulation. Mice are head-fixed to avoid the eye movement evoked by vestibular stimulation. A high-speed camera is used to record the ocular movements from mice viewing moving gratings with changing visual parameters. The physical size of the eyeballs of individual mice is calibrated to ensure the accuracy of deriving the angle of eye movements21. This quantitative method allows comparing OKR behavior between animals of different ages or genetic backgrounds, or monitoring its change caused by pharmacological treatments or visual-motor learning.

Protocol

All experimental procedures performed in this study were approved by the Biological Sciences Local Animal Care Committee, in accordance with guidelines established by the University of Toronto Animal Care Committee and the Canadian Council on Animal Care.

1. Implantation of a head bar on top of the skull

NOTE: To avoid the contribution of VOR behavior to the eye movements, the head of the mouse is immobilized during the OKR test. Therefore, a head bar is surgically implanted on top of the skull.

  1. Anesthetize a mouse (2-5-month-old female and male C57BL/6) by a mixture of 4% isoflurane (v/v) and O2 in a gas chamber. Transfer the mouse to a customized surgery platform and reduce the concentration of isoflurane to 1.5%-2%. Monitor the depth of anesthesia by checking the toe-pinch response and the respiration rate throughout the surgery.
  2. Place a heating pad underneath the animal's body to maintain its body temperature. Apply a layer of lubricant eye ointment to both eyes to protect them from drying. Cover the eyes with aluminum foil to protect them from light illumination.
  3. Subcutaneously inject carprofen at a dose of 20 mg/kg to reduce pain. After wetting the fur with chlorhexidine gluconate skin cleaner, shave the fur on top of the skull. Disinfect the exposed scalp with 70% isopropyl alcohol and chlorhexidine alcohol twice.
  4. Inject bupivacaine (8 mg/kg) subcutaneously at the site of incision, then remove the scalp (~1 cm2) with scissors to expose the dorsal surface of the skull, including the posterior frontal bone, parietal bone, and interparietal bone.
  5. Apply several drops of 1% lidocaine and 1:100,000 epinephrine on the exposed skull to reduce local pain and bleeding. Scrape the skull with a Meyhoefer curette to remove the fascia and clean it with phosphate-buffered saline (PBS).
    NOTE: The temporalis muscle is separated from the skull to increase the surface area for a head bar to attach.
  6. Dry the skull by gently blowing compressed air toward the skull surface until the moisture is gone and the bone turns whitish. Apply a thin layer of superglue on the exposed surface of the skull, including the edge of the cut scalp, followed by a layer of acrylic resin.
    NOTE: The surface of the skull needs to be free of blood or water before the application of superglue.
  7. Place a stainless-steel head bar (see Figure 1A) along the midline on top of the skull. Apply more acrylic resin, starting from the edge of the head bar until the base of the head bar is completely embedded in the acrylic resin. Apply acrylic resin two or three times to build up the thickness.
  8. Wait for about 15 min until the acrylic resin hardens. Subcutaneously inject 1 mL of lactated ringer's solution. Then, return the mouse to a cage placed on a heating pad until the animal is fully mobile.
  9. Allow the mouse to recover in the home cage for at least 5 days after surgery. Once the animal is in good shape, fix its head with the head bar in the OKR setup for 15-30 min to familiarize it with the head fixation and the experimental environment. Repeat the familiarization once a day for at least 3 days.

2. Setup of the virtual drum and video-oculography

  1. Mount three monitors orthogonally to each other to form a square enclosure that covers ~270° of the azimuth and 63° of the elevation in the visual space (Figure 1B left).
  2. With a discrete graphic card, merge the three monitors into a simple display to ensure synchronization across all the monitors.
  3. Calibrate the luminance of monitors as described below.
    1. Turn on the computer to which the monitors are connected and wait for 15 min. The warm-up is essential to have stable luminance.
    2. Systematically change the brightness setting on the monitor from 0 to 100 by steps of 25.
    3. For each brightness value, measure the luminance of the monitors under various pixel values (0-255, steps of 15) with a luminance meter.
    4. Fit the relationship between luminance and brightness for pixel value 255 with linear regression and estimate the brightness value that gives rise to 160 cd/m2.
    5. For each pixel value used in the luminance measurement (step 2.3.3), estimate the luminance for the brightness value derived in step 2.3.4 based on linear regression. Use the power function lum = A * pixelγ to fit the relationship between the new set of luminance values (under the brightness value derived in 2.3.4) and their corresponding pixel values to derive the gamma factor γ and the coefficient A. These will be used to generate sinusoidal gratings of desired luminance values.
    6. Set the brightness of all three monitors to the values derived in step 2.3.4 to ensure their luminance values are the same for the same pixel value.
  4. Generate a virtual drum, which is used to stimulate OKR behavior, with the visual stimulation toolkit, as described below.
    1. Present a vertical sinusoidal grating on the monitors and adjust the period (spacing between stripes) along the azimuth to ensure the projection of the grating on the eye has constant spatial frequency (drum grating; Figure 1B middle and right).
    2. Ensure the animal head-fixed at the center of the enclosure so that it sees the grating has a constant spatial frequency across the surface of the virtual drum.
    3. Modify the parameters of moving grating, such as the oscillatory amplitude, spatial frequency, temporal/oscillation frequency, direction, contrast, etc., in the visual stimulation codes. Use two types of visual motion: (1) the grating drifts clockwise or counterclockwise in an oscillatory manner following a sinusoidal function:
      Equation 1
      Here, Amp is the amplitude of drum trajectory, f is the oscillation frequency, and t is the time (oscillation amplitude: 5°; grating spatial frequency: 0.04-0.45 cpd; oscillation frequency: 0.1-0.8 Hz, corresponding to a peak velocity of the stimulus of 3.14-25.12 °/s [drum velocity = Amp x 2π x f x cos (2π x f x t); contrast: 80%-100%; mean luminance: 35-45 cd/m2; (2) the grating drifts unidirectionally at a constant velocity:
      Equation 2
      ​(Spatial frequency: 0.04-0.64 cpd; temporal frequency: 0.25-1 Hz; drum velocity = temporal frequency/spatial frequency.)
  5. Set up the video-oculography as described below.
    1. To avoid blockage of animal's visual field, place an infrared (IR) mirror 60° from the midline to form an image of the right eye.
    2. Place an IR camera on the right side behind the mouse (Figure 1C left) to capture an image of the right eye.
    3. Mount the high-speed IR camera on a camera arm that allows the camera to rotate by ± 10° around the image of the right eye (Figure 1C right).
    4. Use a photodiode attached to one of the monitors to provide an electrical signal to synchronize the timing of video-oculography and visual stimulation.
    5. Place four IR light emitting diodes (LEDs) supported by gooseneck arms around the right eye to provide IR illumination of the eye.
    6. Place two IR LEDs on the camera to provide corneal reflection (CR) references: one is fixed above the camera (X-CR), while the other is on the left side of the camera (Y-CR; Figure 1D).
    7. Measure the optical magnification of the video-oculography system with a calibration slide.
      NOTE: The reference CRs are used to cancel out the translational eye movements when the eye angle is calculated based on the rotational eye movements.
  6. Fix the animal's head at the center of the enclosure formed by the monitors, as described below.
    1. Fix the animal's head with the head plate to the center of the rig and make it face forward. Adjust the tilt of the head so that the left and right eyes are leveled, and the nasal and temporal corners of the eyes are aligned horizontally (Figure 1E).
    2. Move the animal's head horizontally by coarse adjustment provided by the head-fixation apparatus and fine adjustment provided by a 2D translational stage, and vertically through the head-fixation apparatus and a post/post holder pair, until the animal's right eye appears in the live video of the camera. Before the calibration and measurement of eye movements, overlay the image of the animal's right eye reflected by the hot mirror with the pivot point of the camera arm (see details in step 3.4 below).
  7. Build a customized enclosure around the OKR rig to block room light (Figure 1F).

3. Calibration of eye movements

NOTE: Rotational eye movements are calculated based on movements of the pupil and the radius of the orbit of the pupillary movements (Rp, the distance from the center of the pupil to the center of the eyeball). For each individual mouse, this radius is measured experimentally21.

  1. Fix the animal's head at the center of the enclosure formed by the three monitors, as described in step 2.6.1.
  2. Turn the camera on and adjust the four LEDs surrounding the right eye to achieve uniform IR illumination.
  3. Under visual guidance, adjust the position of the right eye until it appears at the center of the video, as described in step 2.6.2.
  4. Align the virtual image of the right eye with the pivot point of the camera arm, as described below.
    1. Manually rotate the camera arm to the left extreme end (-10°). Manually move the animal's right eye position on the horizontal plane perpendicular to the optical axis with fine adjustment of the 2D translational stage (Figure 1C, green arrow), until the X-CR is at the horizontal center of the image.
    2. Manually rotate the camera arm to the other end (+10°). If the X-CR runs away from the center of the image, move the right eye along the optical axis with fine adjustment until the X-CR comes to the center (Figure 1C, blue arrow).
    3. Repeat steps 3.4.1-3.4.2 a few times until the X-CR stays at the center when the camera arm swings leftward and rightward. If the right eye moves in the middle of one repetition, restart the adjustment process.
  5. Measure the vertical distance between the Y-CR and X-CR after locking the camera arm at the central position. Turn the Y-CR LED on and record its position on the video, and then switch to the X-CR LED and record its position.
    NOTE: The vertical distance between the Y-CR and X-CR will be used to derive the position of the Y-CR during the measurement of eye movements in which only the X-CR LED is turned on.
  6. Measure the radius of pupil rotation Rp, as described below.
    1. Rotate the camera arm to the left end (-10°) and record the positions of the pupil (Pp1) and X-CR (PCR1) on the video.
    2. Then, rotate the camera arm to the right end (+10°) and record the positions of the pupil (Pp2) and X-CR (PCR2) on the video. Repeat this step multiple times.
      NOTE: The animal's right eye needs to remain stationary during each repetition so that the amount of pupil movements in the movie accurately reflects the degree of swinging the camera arm.
    3. Based on the values recorded above, calculate the radius of pupil rotation Rp (Figure 2A) with the following formula:
      Equation 3
      NOTE: The distance between the corneal reflection and pupil center in the physical space is calculated based on their distance in the movie:
      PCR – Pp = number of pixels in the movie x pixel size of camera chip x magnification
  7. Develop the relationship between Rp and pupil diameter, as described below. Rp changes when the pupil dilates or constricts; proximately, its value is inversely proportional to the pupil size (Figure 2B top).
    1. Change the luminance of the monitors systematically from 0 to 160 cd/m2 to regulate the pupil size.
    2. For each luminance value, repeat step 3.6 8-10 times and record the diameter of the pupil.
    3. Apply linear regression to the relationship between Rp and pupil diameter based on the values measured above to derive the slope and intercept (Figure 2B bottom).
      NOTE: The outliers caused by occasional eye movements are removed before the linear fitting. For repetitive measurements in multiple sessions, the calibration needs to be done only once for one animal, unless its eye grows bigger during the experiment.

4. Record eye movements of the OKR

  1. Head-fix a mouse in the rig following steps 3.1-3.4. Skip this step if the recording occurs right after the calibration is done. Lock the camera arm at the central position.
  2. Set up the monitors and the animal for scotopic OKR as described below. Skip this step for photopic OKR.
    1. Cover the screen of each monitor with a customized filter, which is made of five layers of 1.2 neutral density (ND) film. Make sure no light leaks out through the gap between the filter and the monitor.
    2. Turn the room light off. The following steps are done with the aid of an IR goggle.
    3. Apply one drop of pilocarpine solution (2% in saline) to the right eye and wait 15 min. Ensure the drop stays on the eye and is not wiped away by the mouse. If the solution is wiped away by the animal, apply another drop of pilocarpine solution. This shrinks the pupil to a proper size for eye tracking under the scotopic condition.
      NOTE: Under the scotopic condition, the pupil dilates substantially so that its edge is partially hidden behind the eyelid. This affects the precision of estimating the pupil center by video-oculography. Pharmacologically shrinking the pupil of the right eye diminishes its visual input, and thus the visual stimuli are presented to the left eye.
    4. Rinse the right eye with saline to thoroughly wash away the pilocarpine solution. Pull down the curtain to completely seal the enclosure, which prevents stray light from interfering with the scotopic vision.
    5. Give the animal 5 min to fully accommodate to the scotopic environment before starting the OKR test.
  3. Run the visual stimulation software and the eye tracking software. For photopic OKR measurement, ensure the drum grating oscillates horizontally with a sinusoidal trajectory; for scotopic OKR measurement, ensure the drum grating drifts at a constant velocity from left to right, which is the temporo-nasal direction in reference to the left eye.
    NOTE: When the pupil of the right eye, but not of the left eye, is shrunk by pilocarpine under the scotopic condition, the OKR elicited by oscillatory drum stimulation is highly asymmetric. Thus, for scotopic OKR measurement, the left eye is stimulated while the movement of the right eye is monitored.
  4. The eye tracking software automatically measures the pupil size, CR position, and pupil position for each frame, and computes the angle of eye position based on the following formula (Figure 2C):
    Equation 4
    Here, PCR is the CR position, Pp is the pupil position, and Rp is the radius of pupil rotation. The distance between corneal reflection and pupil center in the physical space is calculated based on their distance in the movie:
    PCR – Pp = number of pixels in the movie x pixel size of camera chip x magnification
    Rp of the corresponding pupil size is derived based on the linear regression model in step 3.7.3 (Figure 2B bottom).

5. Analysis of eye movements of the OKR with the eye analysis software

  1. Process the eye traces using a median filter (filter window = 0.05 s) to remove high frequency noise (Figure 3A middle).
  2. Remove the saccades or nystagmus as described below.
    1. Estimate the eye velocity by calculating the first order derivative of eye movements (Figure 3A bottom). Identify the saccades or nystagmus by applying a velocity threshold of 50 °/s (Figure 3A bottom).
    2. Replace the saccades or nystagmus by extrapolating the eye positions during these fast eye movements from the segment before the saccades or nystagmus based on linear regression (Figure 3B).
  3. Compute the amplitude of OKR eye movements by fast Fourier transform (Goertzel algorithm) if the drum grating oscillates (Figure 3C), or compute the average velocity of eye movements during the visual stimulation if the drum grating moves at a constant velocity in one direction (Figure 3B bottom).
    NOTE: The amplitude of oscillatory eye movements derived from Fourier transform is similar to the amplitude derived from the fitting of the eye trajectory with a sinusoidal function (Figure 3D).
  4. Calculate the OKR gain. For oscillatory drum motion, OKR gain is defined as the ratio of the amplitude of eye movements to the amplitude of drum movements (Figure 3C right). For unidirectional drum motion, OKR gain is defined as the ratio of eye velocity to drum grating velocity (Figure 3B bottom).

Representative Results

With the procedure detailed above, we evaluated the dependence of the OKR on several visual features. The example traces shown here were derived using the analysis codes provided in Supplementary Coding File 1, and the example traces raw file can be found in Supplementary Coding File 2. When the drum grating drifted in a sinusoidal trajectory (0.4 Hz), the animal's eye automatically followed the movement of the grating in a similar oscillatory manner (Figure 3B top panel), which is characteristic of the OKR behavior2,5,8. The amplitude of OKR eye movements in the horizontal axis was derived with fast Fourier transform (Figure 3C&D), and the OKR gain was calculated as the ratio of the amplitude of eye movements to the amplitude of grating movement (Figure 3C). The OKR gain varied with the values of spatial frequency, oscillation frequency, and the direction of movement grating (Figure 4A). First, the spatial frequency tuning curve of the OKR behavior had an inverted V-shape and peaked at an intermediate spatial frequency of 0.16 cpd (Figure 4A left). Second, the oscillation frequency tuning curve monotonically declined as the oscillation frequency of the drum grating increased (Figure 4A middle), indicating that the OKR behavior works best in response to the low-speed visual motion4. Both the amplitude and shape of the oscillation frequency tuning curve varied when presenting gratings of different spatial frequencies17. Third, the horizontal OKR could also be induced by gratings moving in different directions (Figure 4A right). The strongest horizontal OKR behavior was elicited by the temporo-nasal motion (0°). The OKR gain dropped to ~80%, or ~30% of the maximum, when the grating moved in oblique angles 30° or 60° deviated from the temporo-nasal direction (both upward and downward), respectively, and the horizontal OKR disappeared when the grating moved vertically upward or downward (90° and 270°). Moreover, the shapes of the tuning curves were influenced by the level of luminance. For example, animals performed the OKR behavior well in response to spatial frequencies of 0.16 and 0.32 cpd under the photopic condition, but the spatial frequency tuning curve shifted leftward under the scotopic condition (Figure 4B). To analyze the shape of tuning curves, we fit them with appropriate mathematical functions. For example, the Gaussian function was used to estimate the peak and bandwidth of spatial frequency tuning (Figure 4C). With this analysis, we found that the tuning curve under the scotopic condition had a lower value in the preferred spatial frequency in comparison to the photopic condition. The procedure detailed above can also be used to quantify the plasticity of OKR behavior. Following 45 min of continuous OKR stimulation, the amplitude of the OKR behavior was significantly potentiated (Figure 4D), consistent with previous reports. These results demonstrate the applications of this protocol in examining the oculomotor behaviors and potentials in understanding the brain circuits involved in these behaviors.

Figure 1
Figure 1: Setup of the OKR rig. (A) Dimensions of the head bar. (B) Back view (left) and top view (middle) of the virtual drum system. Three monitors are mounted orthogonally to each other. The head of a mouse is placed at the center of the square enclosure and faces forward. The period of visual stimulation (space between stripes) is varied based on the azimuth to the animal's eye to ensure the projection of the grating on the eye has a constant spatial frequency. In other words, the spatial frequency of the grating is perceived as constant throughout the visual field, as if the grating drifts along the surface of a virtual drum (right). (C) Setup of IR video-oculography. Left: the position of the camera when fixed at the center during OKR recording. Blue arrow: along the optical axis. Green arrow: perpendicular to the optical axis. Right: the rotation of the camera during eye calibration. (D) Position of the X-CR and Y-CR LEDs fixed on the camera. (E) Levels of the left and right eyes (left), and the nasal and temporal corners of the eyes are aligned horizontally (right) by rotating the horizontal rod or the headplate adapter, respectively (red arrows). (F) Photo of the OKR station. Note that the OKR rig is placed inside a customized enclosure with a black curtain. Abbreviations: IR = infrared; CR = corneal reflection. Please click here to view a larger version of this figure.

Figure 2
Figure 2: Calibration and measurement of eye positions in video-oculography. (A) Schematic of the calibration. The radius of pupil rotation (Rp) is estimated by rotating the camera to the leftmost position (-10°, left panel) and to the rightmost position (10°, right panel). Red dots indicate the positions of corneal reflection X-CR when the camera is placed to the leftmost and rightmost positions. Blue dots indicate the centers of the pupils. Green bars indicate the distances between the corneal reflection and pupil center viewed in the camera video (PCR – PP). (B) Dependence of Rp on pupil size. Top: schematics of eyeballs with a small or large pupil. Bottom: relationship between Rp and the diameter of the pupil of an example mouse. The pupil size is changed by varying the luminance (10 values in the range of 0-160 cd/m2) to the mouse. Black dots: the data that are used for linear fitting. Blue dots: outliers that are excluded from linear fitting. Red curve: the best-fit line in linear regression. Note that the Rp is inversely proportional to the pupil diameter. (C) Calculation of the angle of eye position when the eye has moved to the right or left sides of the optic axis. Red dots, blue dots, and green bars have the same meanings as in A. Please click here to view a larger version of this figure.

Figure 3
Figure 3: Calculation of OKR gain. (A) Top: snapshots of nasal (N; left) and temporal (T; right) eye positions taken during OKR stimulation. Red ellipses: fitting to pupil profile. Red crosses: pupil centers. White arrows: corneal reflection of a reference LED. Middle: the trajectories of eye movements with (black) or without (red) a median filter (filter window = 0.05 s) to remove high-frequency noise. Bottom: estimation of the eye velocity by calculating the first order derivative of eye movements. Saccades (red arrows) are detected with a velocity threshold of 50 °/s. (B) The trajectory of slow eye movements of the OKR after removing the saccades/nystagmus (black) overlaid with drum trajectory22. Top: oscillatory drum motion with an amplitude of 5° and an oscillation frequency of 0.4 Hz. Bottom: unidirectional (temporo-nasal) drum motion with a constant velocity of 6.25 °/s. (C) Left: the cycle-average of the eye trajectory in B top. Right: the frequency analysis of the eye or drum grating movements by fast Fourier transform. It should be noted that the drum grating oscillates at 0.4 Hz, and thus the amplitudes of eye and drum grating movement peak at 0.4 Hz (star marks). OKR gain is the ratio of the amplitudes of eye and drum grating movement at 0.4 Hz. (D) Top: curve fitting of the eye trajectory in B top with sinusoidal function. Bottom: relationship of the amplitude of eye movements derived by the fast Fourier transform method and derived from sinusoidal curve fitting. Yellow dot: example in the top. Please click here to view a larger version of this figure.

Figure 4
Figure 4: The visual feature selectivity and plasticity of OKR behaviors. (A) The visual feature selectivity of OKR gain under the photopic condition. Left: spatial frequency tuning curve from one animal (oscillation frequency: 0.4 Hz; trajectory: oscillating horizontally; mean luminance: 40 cd/m2; n = 15). Middle: oscillation frequency tuning curve from one animal (spatial frequency: 0.08 or 0.16 cpd; trajectory: oscillating horizontally; mean luminance: 40 cd/m2; n = 15). Right: direction tuning curve from one animal (spatial frequency: 0.16 cpd; temporal frequency: 1 Hz; mean luminance: 45 cd/m2; n = 24). The red arrow and bar indicate the temporo-nasal direction. For spatial and temporal/oscillation frequency tuning, a vertical drum grating presented on three monitors moves horizontally at a constant velocity or in an oscillatory manner. For direction tuning, a grating presented on only the right monitor moves in one of 12 directions at a constant velocity. Thickness: standard error of the mean (SEM). (B) Spatial frequency tuning curve of OKR gain from one animal under the scotopic or photopic condition. Photopic: oscillatory motion; oscillation frequency: 0.2 Hz; mean luminance: 40 cd/m2; n = 15. Scotopic: linear motion with a constant velocity; temporal frequency: 0.25 Hz; mean luminance: 8 x 10-5 cd/m2; n = 16. The scotopic condition is achieved by covering the monitors with five layers of Lee filter (299 1.2 ND). Thickness: SEM. (C) Gaussian fitting of spatial frequency tuning of OKR gain under photopic and scotopic conditions. Photopic: oscillatory motion; oscillation frequency: 0.2 Hz; mean luminance: 40 cd/m2; n = 15. Scotopic: linear motion with a constant velocity; temporal frequency: 0.25 Hz; mean luminance: 8 x 10-5 cd/m2; n = 16. (D) OKR potentiation of one mouse induced by 45 min of continuous OKR stimulation. Spatial frequency: 0.1 cpd; oscillation frequency: 0.4 Hz; mean luminance: 35 cd/m2; n = 40. Top: cycle averaged OKR trajectories before and after OKR potentiation. Thickness: SEM. Please click here to view a larger version of this figure.

Supplementary Coding File 1: Analysis code used for generating the example traces. Please click here to download this File.

Supplementary Coding File 2: Example traces generated with the software. Please click here to download this File.

Discussion

The method of the OKR behavioral assay presented here provides several advantages. First, computer-generated visual stimulation solves the intrinsic issues of physical drums. Dealing with the issue that physical drums do not support the systematic examination of spatial frequency, direction, or contrast tuning8, the virtual drum allows these visual parameters to be changed on a trial-by-trial basis, thus facilitating a systematic and quantitative analysis of the feature selectivity of the OKR behavior (Figure 4A); while physical drums suffer from a non-uniform illumination by an external light source23, the virtual drum can readily provide homogeneous luminance across its surface; with the help of ND filters and a luminance meter, the computer-generated visual stimulation allows the OKR measurement at various well-controlled luminance levels from scotopic to photopic (Figure 4B), which is difficult to be done with physical drums. Without the limitation in the acceleration of physical drums due to their inertial mass, the virtual visual stimulation can achieve ideally accurate trajectories, especially at high acceleration and high velocity. Moreover, computer-generated visual stimulation allows the creative design of other types of visual stimulation, such as coherently moving dots, which helps investigate mechanisms of various oculomotor behaviors. Second, our procedure is standardized and thus requires minimal efforts to monitor the progress of behavioral recording, which provides an opportunity to examine multiple mice simultaneously. Therefore, it is suitable for studies involving a large cohort of animals (tens to hundreds of animals). Third, the high precision and quantitative power make it possible to compare repetitive OKR measurements of the same mice in longitudinal studies24, under different pharmacological treatments10, or under neural circuit perturbations5. Last, the Fourier transform-based analysis in the frequency domain5,7,9 gives equivalent results in the amplitude of oscillatory eye movements to the fitting-based analysis in the temporal domain12,25,26 (Figure 3D bottom), demonstrating that the analysis method presented here is both accurate and precise.

Our method also provides an opportunity to study OKR plasticity, a wildly used paradigm to investigate the mechanisms of oculomotor learning. When presenting continuous OKR stimulation to a mouse or surgically lesioning its vestibular organ, the amplitude of the OKR can be potentiated8,9. The OKR assay presented here is sensitive enough to capture small changes in eye movements happening in OKR potentiation (Figure 4D). The quantitative power of this method makes it possible to correlate behavioral changes with neural circuit dynamics, which is essential to reveal the mechanisms underlying oculomotor learning5,8,9,13.

To ensure the accuracy of OKR measurement, there are a few critical steps. First, during the surgery, extra care is needed to avoid superglue and dental acrylic touching the eyelid, which may damage the cornea or partially occlude the eye opening. Second, the strength of the OKR is affected by the behavioral state of mice27,28. Thus, a few rounds of accommodation are recommended to minimize the impact of stress on OKR measurement; also, the disturbance caused by odorants, noise, or light should be prevented during the recording. Finally, the heads of the mice should be oriented properly so that the line connecting the two corners of eyes is parallel to the horizontal axis. This guarantees that the direction of visual motion is aligned with axis of adduction and abduction eye movements. Thirdly, the uniform illumination of the eye is key to generating a sharp image of the pupil and in turn assuring high-quality eye tracking.

It is worth noting that there are a few limitations of the methods presented here. First, when an animal's eye blinks or opaque eye discharge blocks the pupil, the video-oculography loses track of the eye momentarily or permanently. Similarly, it cannot be used to monitor eye movements when the eyelids are sutured. Second, the temporal resolution of the video-oculography is limited by the frame rate of cameras to a range of 4-20 ms. Lastly, the head-fixed preparation does not allow monitoring the ocular behaviors of freely moving animals.

The video-oculography and virtual drum presented here had been successfully applied to characterize the visual feature selectivity and the plasticity of OKR behavior, and to understand the retinal and central circuits involved in mediating and adaptively modulating this behavior. In addition, they can also benefit studies in which other ocular behaviors are either the topics or even confounding factors of the neural phenomena. For example, the video-oculography can be used to monitor pupil dilation29 and saccade-like eye movements30,31, which are indicative of alertness and brain state32,33,34,35. Moreover, the calibration and measurement procedures outlined here are universally applicable for monitoring eye movements with a head-mounted camera in freely moving mice.

Disclosures

The authors have nothing to disclose.

Acknowledgements

We are thankful to Yingtian He for sharing data of direction tuning. This work was supported by grants from the Canadian Foundation of Innovation and Ontario Research Fund (CFI/ORF project no. 37597), NSERC (RGPIN-2019-06479), CIHR (Project Grant 437007), and Connaught New Researcher Awards.

Materials

2D translational stage Thorlabs XYT1
Acrylic resin Lang Dental B1356 For fixing headplate on skull and protecting skull
Bupivacaine STERIMAX ST-BX223 Bupivacaine Injection BP 0.5%. Local anesthesia
Carprofen RIMADYL 8507-14-1 Analgesia
Compressed air Dust-Off
Eye ointment Alcon Systane For maintaining moisture of eyes
Graphic card NVIDIA Geforce GTX 1650 or Quadro P620. For generating single screen among three monitors
Heating pad Kent Scientific HTP-1500 For maintaining body temperature
High-speed infrared (IR) camera Teledyne Dalsa G3-GM12-M0640 For recording eye rotation
IR LED Digikey PDI-E803-ND For CR reference and the illumination of the eye
IR mirror Edmund optics 64-471 For reflecting image of eye
Isoflurane FRESENIUS KABI CP0406V2
Labview National instruments version 2014 eye tracking
Lactated ringer BAXTER JB2324 Water and energy supply
Lidocaine and epinephrine mix Dentsply Sirona 82215-1 XYLOCAINE. Local anesthesia
Luminance Meter Konica Minolta LS-150 for calibration of monitors
Matlab MathWorks version xxx analysis of eye movements
Meyhoefer Curette World Precision Instruments 501773 For scraping skull and removing fascia
Microscope calibration slide Amscope MR095 to measure the magnification of video-oculography
Monitors Acer  B247W Visual stimulation
Neutral density filter Lee filters 299 to generate scotopic visual stimulation
Nigh vision goggle Alpha optics AO-3277 for scotopic OKR
Photodiode Digikey TSL254-R-LF-ND to synchronize visual stimulation and video-oculography
Pilocarpine hydrochloride Sigma-Aldrich P6503
Post Thorlabs TR1.5
Post holder Thorlabs PH1
PsychoPy open source software version xxx visual stimulation toolkit
Scissor RWD S12003-09 For skin removal
Superglue Krazy Glue Type: All purpose. For adhering headplate on the skull

References

  1. Gerhard, D. Neuroscience. 5th Edition. Yale Journal of Biology and Medicine. , (2013).
  2. Distler, C., Hoffmann, K. P. . The Oxford Handbook of Eye Movement. , 65-83 (2011).
  3. Sereno, A. B., Bolding, M. S. . Executive Functions: Eye Movements and Human Neurological Disorders. , (2017).
  4. Giolli, R. A., Blanks, R. H. I., Lui, F. The accessory optic system: basic organization with an update on connectivity, neurochemistry, and function. Progress in Brain Research. 151, 407-440 (2006).
  5. Liu, B. H., Huberman, A. D., Scanziani, M. Cortico-fugal output from visual cortex promotes plasticity of innate motor behaviour. Nature. 538 (7625), 383-387 (2016).
  6. Prusky, G. T., Alam, N. M., Beekman, S., Douglas, R. M. Rapid quantification of adult and developing mouse spatial vision using a virtual optomotor system. Investigative Ophthalmology & Visual Science. 45 (12), 4611-4616 (2004).
  7. Stahl, J. S., van Alphen, A. M., De Zeeuw, C. I. A comparison of video and magnetic search coil recordings of mouse eye movements. Journal of Neuroscience Methods. 99 (1-2), 101-110 (2000).
  8. Faulstich, B. M., Onori, K. A., du Lac, S. Comparison of plasticity and development of mouse optokinetic and vestibulo-ocular reflexes suggests differential gain control mechanisms. Vision Research. 44 (28), 3419-3427 (2004).
  9. Katoh, A., Kitazawa, H., Itohara, S., Nagao, S. Dynamic characteristics and adaptability of mouse vestibulo-ocular and optokinetic response eye movements and the role of the flocculo-olivary system revealed by chemical lesions. Proceedings of the National Academy of Sciences. 95 (13), 7705-7710 (1998).
  10. Cahill, H., Nathans, J. The optokinetic reflex as a tool for quantitative analyses of nervous system function in mice: application to genetic and drug-induced variation. PLoS One. 3 (4), 2055 (2008).
  11. Cameron, D. J., et al. The optokinetic response as a quantitative measure of visual acuity in zebrafish. Journal of Visualized Experiments. (80), 50832 (2013).
  12. de Jeu, M., De Zeeuw, C. I. Video-oculography in mice. Journal of Visualized Experiments. (65), e3971 (2012).
  13. Kodama, T., du Lac, S. Adaptive acceleration of visually evoked smooth eye movements in mice. The Journal of Neuroscience. 36 (25), 6836-6849 (2016).
  14. Doering, C. J., et al. Modified Ca(v)1.4 expression in the Cacna1f(nob2) mouse due to alternative splicing of an ETn inserted in exon 2. PLoS One. 3 (7), e2538 (2008).
  15. Shi, C., et al. Optimization of optomotor response-based visual function assessment in mice. Scientific Reports. 8 (1), 9708 (2018).
  16. Waldner, D. M., et al. Transgenic expression of Cacna1f rescues vision and retinal morphology in a mouse model of congenital stationary night blindness 2A (CSNB2A). Translational Vision Science & Technology. 9 (11), 19 (2020).
  17. Tabata, H., Shimizu, N., Wada, Y., Miura, K., Kawano, K. Initiation of the optokinetic response (OKR) in mice. Journal of Vision. 10 (1), 1-17 (2010).
  18. Al-Khindi, T., et al. The transcription factor Tbx5 regulates direction-selective retinal ganglion cell development and image stabilization. Current Biology. 32 (19), 4286-4298 (2022).
  19. Harris, S. C., Dunn, F. A. Asymmetric retinal direction tuning predicts optokinetic eye movements across stimulus conditions. eLife. 12, e81780 (2023).
  20. van Alphen, B., Winkelman, B. H., Frens, M. A. Three-dimensional optokinetic eye movements in the C57BL/6J mouse. Investigative Ophthalmology & Visual Science. 51 (1), 623-630 (2010).
  21. Stahl, J. S. Calcium channelopathy mutants and their role in ocular motor research. Annals of the New York Academy of Sciences. 956, 64-74 (2002).
  22. Endo, S., et al. Dual involvement of G-substrate in motor learning revealed by gene deletion. Proceedings of the National Academy of Sciences. 106 (9), 3525-3530 (2009).
  23. Thomas, B. B., Seiler, M. J., Sadda, S. R., Coffey, P. J., Aramant, R. B. Optokinetic test to evaluate visual acuity of each eye independently. Journal of Neuroscience Methods. 138 (1-2), 7-13 (2004).
  24. Burroughs, S. L., Kaja, S., Koulen, P. Quantification of deficits in spatial visual function of mouse models for glaucoma. Investigative Ophthalmology & Visual Science. 52 (6), 3654-3659 (2011).
  25. Wakita, R., et al. Differential regulations of vestibulo-ocular reflex and optokinetic response by β- and α2-adrenergic receptors in the cerebellar flocculus. Scientific Reports. 7 (1), 3944 (2017).
  26. Dehmelt, F. A., et al. Spherical arena reveals optokinetic response tuning to stimulus location, size, and frequency across entire visual field of larval zebrafish. eLife. 10, e63355 (2021).
  27. Magnusson, M., Pyykko, I., Jantti, V. Effect of alertness and visual attention on optokinetic nystagmus in humans. American Journal of Otolaryngology. 6 (6), 419-425 (1985).
  28. Collins, W. E., Schroeder, D. J., Elam, G. W. Effects of D-amphetamine and of secobarbital on optokinetic and rotation-induced nystagmus. Aviation, Space, and Environmental Medicine. 46 (4), 357-364 (1975).
  29. Reimer, J., et al. Pupil fluctuations track fast switching of cortical states during quiet wakefulness. Neuron. 84 (2), 355-362 (2014).
  30. Sakatani, T., Isa, T. PC-based high-speed video-oculography for measuring rapid eye movements in mice. Neuroscience Research. 49 (1), 123-131 (2004).
  31. Sakatani, T., Isa, T. Quantitative analysis of spontaneous saccade-like rapid eye movements in C57BL/6 mice. Neuroscience Research. 58 (3), 324-331 (2007).
  32. Vinck, M., Batista-Brito, R., Knoblich, U., Cardin, J. A. Arousal and locomotion make distinct contributions to cortical activity patterns and visual encoding. Neuron. 86 (3), 740-754 (2015).
  33. Bradley, M. M., Miccoli, L., Escrig, M. A., Lang, P. J. The pupil as a measure of emotional arousal and autonomic activation. Psychophysiology. 45 (4), 602-607 (2008).
  34. Hess, E. H., Polt, J. M. Pupil size as related to interest value of visual stimuli. Science. 132 (3423), 349-350 (1960).
  35. Di Stasi, L. L., Catena, A., Canas, J. J., Macknik, S. L., Martinez-Conde, S. Saccadic velocity as an arousal index in naturalistic tasks. Neuroscience and Biobehavioral Reviews. 37 (5), 968-975 (2013).

Play Video

Cite This Article
Liu, J., Liu, B. Quantification of Visual Feature Selectivity of the Optokinetic Reflex in Mice. J. Vis. Exp. (196), e65281, doi:10.3791/65281 (2023).

View Video