Abstract
Spatial cognition research requires behavioral paradigms that can distinguish between different navigational elements, such as allocentric (map-like) navigation and egocentric (e.g., body centered) navigation. To fill this need, we developed a flexible experimental platform that can be quickly modified without the need for significant changes to software and hardware. In this paper, we present this inexpensive and flexible behavioral platform paired with software which we are making freely available. Our behavioral platform serves as the foundation for a range of experiments, and although developed for assessing spatial cognition, it also has applications in the nonspatial domain of behavioral testing. There are two components of the software platform, “Maze” and “Stim Trigger.” While intended as a general platform, presently both programs can work in conjunction with Neuralynx and Open Ephys electrophysiology acquisition systems, allowing for precise time stamping of neural events. The Maze program includes functionality for automatic reward delivery based on user defined zones. “Stim Trigger” permits control of brain stimulation via any equipment that can be paired with an Arduino board. We seek to share our software and leverage the potential by expanding functionality in the future to meet the needs of a larger community of researchers.
Significance Statement
This paper presents an innovative and cost-effective behavioral platform designed to distinguish between different navigational elements, addressing the crucial need for better spatial cognition research paradigms. The platform’s flexibility allows for quick modifications without major software or hardware changes. Additionally, the freely available software, comprising “Maze” and “Stim Trigger” components, enables precise timestamping of neural events with behavior, while facilitating automatic reward delivery and brain stimulation control. Beyond spatial cognition assessment, the platform’s adaptability extends to nonspatial behavioral testing. The authors’ intention in openly sharing this software is to stimulate collaboration, spur future advancements, and extend its utility to a wider community of researchers. This platform represents a significant advancement in spatial cognition research and behavioral experimentation methods.
Introduction
Spatial cognition is a burgeoning research field in neuroscience that has implications in memory, aging, neurodevelopmental disorders, and progressive neurologic disorders. In the 1940s, Edward C. Tolman’s pioneering experiments revolutionized our comprehension of spatial learning and cognitive mapping. Through innovative research with rats, Tolman introduced the concept of cognitive maps, intricate mental representations enabling animals to navigate and comprehend their environments (Tolman, 1948). Following his pioneering work, over the past 70 years landmark experiments like the “Morris water maze” developed by Morris, illuminated the role of spatial memory in rodents (Morris, 1981). Similarly, O’Keefe and Dostrovsky’s discovery of “place cells” in the hippocampus (O’Keefe and Dostrovsky, 1971) laying the foundation for understanding neural mechanisms underlying spatial navigation. Subsequently, researchers identified “head direction cells” (Taube, 1995) and “grid cells” (Fyhn et al., 2004), further enriching our knowledge of spatial cognition. These foundational studies continue to shape current spatial cognition research, inspiring innovative approaches to investigating memory, learning, and cognitive mapping in both animals and humans. Today, advancements in virtual reality technology empower researchers to simulate intricate real-world environments, revolutionizing our understanding of spatial navigation (Moffat et al., 1998). Moreover, technology advancement has allowed the implementation of complex experimental designs that help with dissecting the innerworkings of spatial navigation behavior and its neural basis (Ormond and O’Keefe, 2022). These innovative approaches have enabled investigations into how learning and memory, aging, and brain disease or disorders intersect with spatial cognition, fostering groundbreaking insights (Cogné et al., 2017). Spatial memory is present in most animal species, and it is a key element in how individuals interact with their surroundings (Benhamou and Poucet, 1995). The significance of spatial navigation across species can be observed in the wide variety of strategies that different species showcase, such as individual approaches, like those observed in rats and solitarily foraging ants, and group-based strategies like the large-scale migratory behaviors of sea turtles and lepidopterans (Freas and Cheng, 2022). The use of animal models is common because they provide methodological benefits, such as the ability to use invasive electrophysiological recording or imaging devices in pertinent brain regions.
The use of animal models in research requires the creation of innovative tools that enable scientists to investigate new concepts related to spatial navigation, learning and memory, anxiety-like behavior, decision-making, and other behavioral variables of interest. Researchers commonly employ mazes to assess spatial cognition, learning, and memory (Paul et al., 2009). These include the radial arm maze, T and Y -mazes, Morris water maze, and Barnes maze, which are among the most frequently used (Gawel et al., 2019). The design of each maze and the task’s protocol dictate which aspects of spatial memory and cognition are evaluated.
The Morris water maze and Barnes maze are examples of open mazes that offer multiple paths to navigate toward a designated objective, typically a shelter or escape platform, using the cues surrounding the testing area (Morris, 1981; Barnes, 1988). The task-learning process for these mazes generally relies on negative reinforcers, in which the removal or avoidance of an aversive stimulus increases the likelihood of a desired behavior recurring. In the Morris water maze, the stress induced is greater than that induced by the Barnes maze, as measured by plasma corticosterone levels. (Harrison et al., 2009). Alternatively, the radial arm maze, the T and Y- mazes, as well as versions such as the Cincinnati maze offer restricted route options (Paul et al., 2009). These mazes have led to significant advances in the field, for example, place cells were first observed in a restricted-route maze (O’Keefe, 1976). Restricted route mazes are often used to study working memory and typically use positive reinforcers like food and water to motivate behavior (Dudchenko, 2004; Morellini, 2013). Albeit with some exceptions, the afore mentioned mazes focus on allocentric (map-like) processing of space. Allocentric processing relies on distal cues or “landmarks” to create a spatial reference frame (Bermudez-Contreras et al., 2020). A remarkable exception is the study published by Rondi-Reig and colleagues (2006), showing that their “starmaze” could be used to develop allocentric, and sequential-egocentric tasks (Rondi-Reig et al., 2006). Furthermore, studies on animal models, such as rodents navigating mazes, have shown that they rely more on local cues and directional signals to guide movements, which is consistent with stimulus-response learning rather than true egocentric navigation based on self-positioning within the environment (Tolman, 1948; Geerts et al., 2020). Thus, there is a need for experimental designs that further investigate the egocentric (e.g., body centered) frame as well as the interaction between both allocentric and egocentric frames of reference.
We developed a new platform to fill a need for flexibility to advance spatial cognition research using novel experimental designs and apparatuses including those that go beyond allocentric spatial location processing. It is common practice to modify mazes and protocols, as seen in combinations like a radial arm maze submerged in a water tank (Buresová et al., 1985). That is why we created a cost-efficient maze with a design based on a circular platform, similar to those used in the Barnes maze, which, when combined with custom MATLAB software, enables researchers to quickly develop customized experimental designs both within the field of spatial cognition research and beyond. Further, this new platform can be used to automate aspects of standerd existing mazes such as the Barnes, T and Y -mazes.
Recording the activity of neurons in freely moving animals during task performance provides valuable insight into how different areas of the brain communicate with each other. The brain dynamics between areas such as the hippocampus, entorhinal cortex, subiculum, thalamus, retrosplenial, and parietal cortex are known to be critical for spatial cognition and other high-level cognitive processes like learning, memory, decision-making, and attention (Bermudez-Contreras et al., 2020). To record neuronal activity simultaneously, researchers use recording arrays with tetrodes or multichannel silicon probes. Additionally, silicone probes offer the ability to record interactions within layered structures more easily, which is essential for understanding brain computations (Buzsáki et al., 2003).
Since the first publication using this software platform with tetrodes (Bower et al., 2005), we have developed and integrated the software to facilitate in vivo electrophysiological experiments using several recording platforms. The tasks that have employed it so far are focused on spatial cognition; however, the platform is flexible enough to allow for testing using many paradigms such as decision-making. To achieve our goal, we implemented two commonly used electrophysiology recording platforms, Neuralynx and Open Ephys, along with two real-time video tracking approaches. We chose these platforms because they offer complementary features. Neuralynx is widely used for its excellent analog signal processing capabilities (though newer systems are capable of analog and digital processing), while Open Ephys provides a cost-effective and open-source solution for digital signal processing. Accurate interpretation of the behavioral correlates of brain signals depends on precise synchronization of video and brain signals. Neuralynx includes the necessary video tracking features, while Open Ephys can be paired with other platforms such as Bonsai, an open-source software program for real-time video analysis and more. Furthermore, our platform has built in functionality for optogenetic experiments, which can modulate brain activity with high specificity using Transistor-Transistor-Logic binary signals, generated by turning a transistor on or off (TTL), signals to trigger laser pulses at specific user defined points in the maze.
Our platform provides the hardware and software components necessary to flexibly and precisely measure brain signals during custom-designed tasks in freely moving animals if dictated by the experimental design or to conduct behavioral experiments without recording brain activity. Furthermore, the prospect exists for combining it with in vivo brain imaging techniques, such as the UCLA Miniscope, although this has not yet been realized.
Materials and Methods
Hardware
Platform
The circular platform (T-60RT, Unfinished Furniture of Wilmington) is a 60-inch diameter Parawood tabletop, 3/4 inch thick. To seal the surface and improve paint adhesion we used KILZ water-based acrylic primer. The paint used was Chalky Finish Krylon “Misty Gray.” The color was selected to provide a neutral background to offer contrast with the rodents’ coat color, which in our case is white, brown, or black for video tracking purposes. However, other colors can be used to increase this contrast for other combinations of rodent strains or to mimic the color of the proximal or distal walls.
The wall is made of low-density polyethylene and has a dual function: it holds the electronic components at a proper height for the animals to interact with them, and serves as a low boundary for the animals without obstructing the view of the surroundings, helping them use distal cues and avoid unintentionally crossing the edge of the maze. Different wall heights and other surface shapes can be used to assess the role of boundaries in the hippocampal processing of space, as observed in the activity of “boundary cells” (Hartley et al., 2000; Solstad et al., 2008). Our software and hardware platform can work with any maze shape or apparatus. Moreover, wall color and height can be manipulated to investigate the impact of its use as a proximal cue. For example, by matching the color of the floor, the effect of walls as a proximal cue can be further reduced. Furthermore, the walls can be removed if the platform is placed high enough to reduce the chances of an animal jumping off (Gawel et al., 2019).
LED cue lights
We installed 32 evenly spaced CHANZON LED lights (3 V, 20 mA) on the maze wall using eBoot JST SM 2 Pin Plugs. In our configuration, each diode is connected to a 1.1-K resistor to reduce its brightness. Different resistors can be used to adjust the LEDs to the desired brightness.
The wiring of electronic components, including the LEDs, uses XHF lever connectors for safety and easy modification. The LEDs are connected to a 48-channel USB-DIO-48 module through a Sysly(c) IDC50-B breakout board (Fig. 1).
Pinout mapping for IDC breakout boards in the behavioral platform connecting the electronic components to USB Access I/O modules. Each pin offers a connection to control LEDs, solenoids, or send TTL signals. The circle denotes the maze surface and which electronic components are mounted under the maze.
Our Maze software (MATLAB) controls 48 TTL-compatible bits in the I/O module using the AIOUSB API library (ACCES I/O, 2022). The LED lights can be used as cues during experimental procedures and to automatically calibrate the maze location and zones of interest in the camera frame.
Hardware systems
Liquid reward
To deliver liquid rewards to the animals, we incorporated eight solenoid-operated pinch valves (Cole-Parmer), which are linked to tubes containing the liquid. These valves facilitate the dispensing of water or nutritional drinks (such as Ensure) to animals undergoing water or food restriction, respectively. The tubes are connected to uniformly spaced spouts on the maze walls, with one spout positioned over every fourth LED. The use of fewer spouts would not necessitate any modifications to the maze software.
These valves are controlled using a 16-channel Digital I/O USB Module (USB-IDO-16; ACCES I/O, 2022). The Maze software (MATLAB) uses the afore-mentioned AIOUSB API library to control the delivery of the liquid (ACCES I/O, 2022).
Brain stimulation or Optogenetic stimulation
Stimulation of the medial forebrain bundle is a strong reinforcer that is extensively used in animal models (Olds and Milner, 1954; Wise and Rompre, 1989; Wise, 1996, 2002). Various laboratories have successfully used brain stimulation as a reinforcer in spatial cognition tasks in both mice and rats (Euston et al., 2007; Wilber et al., 2014, 2017; Stimmell et al., 2019; Cushing et al., 2020). Further, optogenetics is a powerful tool for modulating brain activity with temporal and spatial specificity (Emiliani et al., 2022).
We provide two options for signal to trigger stimulation: unipolar or bipolar. We use a TTL signal designed to trigger a bipolar output by a constant current stimulus isolator (SYS-A365, World Precision Instruments). A data acquisition device (USB-1208FS-Plus, Measurement Computing) gates the TTL pulses used to control stimulation parameters via MATLAB using the MATLAB Data Acquisition Toolbox. The characteristics of the stimulation, such as frequency, duration and duty cycle can be determined in two separate software packages: “Stim Trigger” (for testing stimulation parameters alone in an operant chamber) and “Maze” (for delivering brain stimulation during experiments on the maze platform; MATLAB Runtime 9.2). For further details on selecting stimulation parameters for rodents, see Carlezon and Chartoff (2007). Duty cycle determines the percentage of each stimulus cycle that is set as positive however the number of stimuli depends on the duration and frequency.
The unipolar TTL signal from the data acquisition device can be used as input for optogenetic laser control. The device can connect to the power supply for the laser via BNC (“Bayonet Neill–Concelman”) which is a quick connect/disconnect connector used for coaxial cables. The signal features can be determined in our software similarly to those for electrical stimulation.
Electrophysiology data acquisition
Headstages and probes
We have used a variety of configurations for silicon probes. However, we will focus on the mice configuration using a single, H5 probe (Cambridge Neurotech). We attached the probe to an implantable nano-Drive. The probe was connected to a miniamplifier via an interposer board (Cambridge Neurotech). We used a custom-made cable (Neuralynx), containing SPI and MDR50 male connectors to connect the amplifier to a commutator (Saturn, Neuralynx) which helped to reduce torsion from the freely moving animal and safeguard the integrity of the components. We used a second custom-cable to connect the commutator to either the Open Ephys Acquisition Board or the Neuralynx Digital Lynx SX acquisition system.
For tetrode recording, we used recording arrays built in-house that consisted of 18 or more independently drivable tetrodes for rats (Kloosterman et al., 2009; Wilber et al., 2017) or 16 independently driveable tetrodes for mice (Cushing et al., 2020; Open Ephys and Contributors, 2022). We have used a variety of electrode interface boards to interface with Neuralynx Digital Lynx SX: EIB-72-QC-Small for mice, and EIB-72-QC-Large, EIB-27-18TT, EIB-36-16TT, HS-72-QC-LED for rat recordings (Neuralynx).
Acquisition systems
The Open Ephys Acquisition Board and the Digital Lynx SX by Neuralynx are two popular systems for acquiring extracellular electrophysiology data. The Open Ephys board offers an interface between up to 512 channels of data and the computer via USB connection. As the project is open source, components and assembly instructions are freely available. Alternatively, pre-assembled components and remote training can be purchased (Open Ephys and Contributors, 2022).
The Digital Lynx SX offers up to 512 channels with a modular configuration that can be customized to meet the needs and budget of experimenters. Additionally, it includes a Hardware Processing Platform for real-time closed loop data processing, and other advanced experiment designs using platforms such as MATLAB (Neuralynx, 2002).
Both systems offer a range of features and flexibility, making them popular choices for researchers conducting electrophysiology experiments.
Software
Our software is available in an OSF repository at: https://osf.io/svtzr/.
Maze
In developing our maze software, we used the App Designer platform, a visual environment in MATLAB that enables the rapid creation of graphical user interfaces. This choice provides users with the ability to modify our code package in a relatively straightforward manner to better suit their needs. Some examples of the use of this software with further explanation of the functionalities used can be found in Results. The GUI includes the following functionalities (Fig. 2).
“Maze” GUI. A, Control of electronic hardware including reward type, stimulation parameters, and valve control. B, Characteristics of zones of interest, including number of zones. C, Counters and configuration controls. D, Zone location, selection of “Zones” and “Rewarded zones,” and randomization of rewarded zones. E, LED test and auto calibration controls. F, Animal ID and comment boxes. G, Real-time position tracking. Zones are marked with * along with their identification number.
Control of electronic hardware
Users can control all electronic hardware through the GUI. The control for the valves becomes available when users select Liquid reward (Fig. 2A). If users select Brain stimulation, the panel with the stimulation settings becomes visible and editable, along with a counter for manual (user-triggered) and animal-triggered stimulation. The settings that can be modified include duration, frequency, pulse profile, and delay after entry to the rewarded zone (this feature requires the animal to remain in the reward zone for the experimenter specified delay in order for the animal to obtain a stimulation).
In the “Zone coordinates” panel (Fig. 2D), buttons corresponding to each zone number allow you to turn the LEDs on and off. When turned on they will become active when the zone is the next/active in the sequence. The LED can be set to blink (or not) and the blink rate can be specified. All changes to the LED status are recorded and timestamped in the acquisition system via TTL sent from the I/O board.
Set zone characteristics
The software is designed around spatial zones that are used to record maze events (e.g., when the animal crosses a particular region of the maze) or trigger rewards (e.g., when the animal reaches a specific location). The GUI allows users to set the characteristics of zones of interest (Fig. 2B). Users can adjust the number of zones to be tracked, and each time an animal enters or exits a zone, the zone number (1–32) is sent via TTL in binary format to the neural signal recording system via the USB-DIO-48 board where the event is precisely timestamped. The zone radius and hysteresis can also be adjusted. Zone hysteresis temporarily increases the zone radius when an animal enters a zone, reducing the likelihood of detecting slight head movements as false zone exits. When the hysteresis option is used, once the animal leaves the zone or the reward is delivered, the zone radius returns to its previous value. The “Zones coordinates” panel (Fig. 2D) displays the numbers and camera pixel coordinates of up to 32 possible zones, as well as LED control buttons. To determine the location of each zone, users have two options: they can manually enter the xy pixel coordinates into the input boxes, or they can use the automatic calibration system (Fig. 2E). With the latter option, the system detects the location of each zone-associated LED using either Cheetah or Bonsai (Fig. 2A). The LEDs are detected sequentially as they light up from 1 to 32, and the system automatically places a zone at each detected location. The IDs for the zones of interest can be entered in the “Zones” panel (Fig. 2D). The “Zones” are indicated in green font to differentiate them from the inactive zones shown in black (Fig. 2G). Associating specific LEDs with specific zones allows the user to illuminate those LEDs as cues for the animal, either on every trial or only on user-defined trials. Additionally, cue lights can be positioned out of the view of the animal so that the experimenter can use them to ensure the maze software is tracking the animal’s sequential progress through the maze accurately, or the LEDs can be turned off and not used at all.
Zone-triggered rewards
The GUI offers zone-triggered reward functionality. The number of rewarded zones must align with the zone IDs entered under the “Rewarded Zones” panel (Fig. 2D), and these zones must also be included in the “Zones.” At any given time, only one zone can be set to be rewarded (i.e., until the animal reaches that zone, or the experimenter manually advances the animal to the next reward zone, the subsequent reward zone, if specified, remains inactive). Once the animal or a manual zone trigger occurs, the next zone in the “Rewarded Zone” panel becomes active. The currently rewarded zone is visually marked with a blue encirclement in the “Animal tracking” panel (Fig. 2G). Reward zones can be repeated and set to activate sequentially (if multiple reward zones are in use). Randomizing the selection order of these zones is also an option. Lists can contain a mix of rewarded and unrewarded zones, allowing some zones to track and timestamp the animal’s progress through key maze segments, while others can be timestamped and automatically paired with a reward if the zone’s reward criteria are met. Users can determine the type of reward bestowed on the animal on entry into one of the active rewarded zones (up to 32 zones can be active for an experiment) by using the “Reward Type” selection buttons.
Real-time tracking of animal’s position
The GUI displays real-time tracking of the animal’s position in relation to user-defined maze zones of interest (Fig. 2G). In the bottom right section of the GUI, the “Animal tracking” panel takes up most of the available space. The tracker displays the location of the animal as a red dot on the axes and provides numerical values for the “X” and “Y” coordinates. If a two-color (e.g., LED) tracking system is set up, and HD tracking is enabled in either Cheetah or Bonsai, the angle of the animal’s head is also displayed in the “Angle” text box (Fig. 2F). Otherwise, the heading is recorded/displayed as 0 degrees. The head direction is visualized as a blue tangent line that shows the orientation of the head.
Animal ID and comment
The right section of the GUI contains the Animal ID and Comment text boxes (Fig. 2F). The contents of these boxes are saved as a variable in the settings file when a session is ended or when the “Save Config File” button is pressed. This allows the complete configuration of all parameters to be loaded from the same file and stored as a record of the experiment details for that session.
Settings file
All the configuration variables are saved in settings files created and loaded in the program (Fig. 2C). This can be used to quickly store and load different experimental protocols to ensure consistency across days. The settings file includes the zone locations, reward variable, animal’s ID and notes from the loaded session. The settings files are stored in .mat format and contain all the settings as variables that can be conveniently accessed with MATLAB and read in with analysis code without the need to run the Maze software.
Stim Trigger
“Stim Trigger” provides a simplified GUI for controlling brain stimulation parameters and triggering methods, along with manual stimulation options and a counter to keep track of the number of delivered stimulations without the need of the full featured maze software (Fig. 3).
“Stim Trigger” software GUI. Top, Stimulus setting input boxes. Middle, “Send Manually” triggers stimulation with each mouse click, “Start Ext Trigger” turns on the automatic trigger, “Clear Stim” resets the stimulation counters (right). Bottom, Radio Buttons to select the timing of the automatic trigger on the rising edge or the falling edge of the input TTL from the external triggering source.
Its main purpose is to offer an easy-to-use platform for researchers to control brain stimulation and integrate it with other behavioral equipment or software. The software provides control over electrical or optogenetic brain stimulation, allowing for both manual triggering by the experimenter and automatic triggering by the animal. This stimulation is converted into a bipolar stimulus by the stimulus isolator. Alternatively, a unipolar TTL signal can be used to regulate a constant or pulsed laser output for optogenetics. The former method is used alongside a nose poke that operates with a +5-V TTL. Moreover, a Virtual Reality maze is integrated, using tablets and an Arduino, to deliver electrical brain stimulation rewards when the animal occupies a specific location in the virtual environment (Molina et al., 2016; Arduino, 2022). Manually triggered stimulation can also be employed to shape behavior, such as training animals to approach the nose poke. The software empowers users to configure stimulation parameters through the GUI and utilizes the data acquisition device to trigger the stimulus isolator. As explained above, the duration and frequency determine the main characteristics of the unipolar signal and the duty cycle represents the ratio of time the stimulus is on compared with the time the stimulus is off.
Bonsai
Bonsai is a powerful open-source software designed for processing heterogeneous streams of data (https://bonsai-rx.org). It is particularly well-suited for real-time video analysis because of its advanced features and flexibility. In our case, we used LEDs mounted on the recording array to determine the animal’s location and then transferred the location via a UDP port to MATLAB (Fig. 4). Additionally, Bonsai also allows for use of visual methods to determine the animal’s position that do not rely on LEDs, as long as the animal is distinguishable from the background.
Bonsai workflow for animal’s location extraction and transfer to Maze. Each graphical element represents a function within the data processing pipeline. From left to right, The live video is processed to extract and transmit the animal’s location to both Maze and a .csv file.
The animal’s location is extracted using various functions provided by Bonsai, such as Bonsai.Vision.Threshold, Bonsai.Vision.FindContours, Bonsai.Vision.BinaryRegionAnalysis, Bonsai.Vision.LargestBinaryRegion, and MemberSelector.Centroid. To standardize for different frame sizes the location of the animal is normalized using “Size” from the Video Camera frame. The information extracted by this pipeline is sent using Bonsai.Osc.SendMessage using the UDP port created via Bonsai.Osc.CreateUdpClient.
The threshold settings must be changed to use Bonsai for automatic calibration using the LEDs in the platform. In the presented workflow, we create a .csv file to save the number of frames and timestamps. The location of the animal can also be saved in this file to have a log for the timestamped animal’s location.
Data availability
The presented code, except analysis code, is available in our data repository in OSF, link: https://osf.io/svtzr/. The list of components to build the platform is available on Google Sheets, link: https://docs.google.com/spreadsheets/d/1GMOYG8HO3yyJB4AWzgjLCo-qlfUHs2DzC6-nCg6M2rA/edit?usp=sharing.
Results
To demonstrate the potential of our platform we will describe a number of applications for these software packages across three laboratories and four institutions. The first will be of behavioral data collected from a “spatial reorientation” task that requires mice to use distal cues to get reoriented in space while running down a linear track (Stimmell et al., 2019). Furthermore, we would like to highlight additional publications that have used this platform. These include three publications employing a sequence task in rats (Bower et al., 2005; Euston and McNaughton, 2006; Euston et al., 2007), one publication using a task that requires rats to navigate toward a randomly selected cue light out of 32 possible options (Wilber et al., 2014), and one using both (Wilber et al., 2017). It is worth noting that all of these publications, involving both the sequence task and the random lights task, incorporated a pre-training phase training rats to shuttle between the ends of a linear track. Additionally, we present an instance where the “Stim Trigger” software was combined with a virtual maze to conduct experiments with mice (Cushing et al., 2020). We also include an example with the object-place paired associate (OPPA) task (Sanchez et al., 2019). Finally, we describe its implementation on an unpublished task, the map-to-action transformation (MAT) task.
First, we will elaborate on the application of this software platform for the spatial reorientation task, which was previously described, and has been used for both rats and mice (Rosenzweig et al., 2003; Stimmell et al., 2019). Briefly, the animal must travel back and forth over a linear track with an unmarked rewarded location. The starting point within the linear track is randomized, thus the distance to the rewarded location changes across trials. The animal is moved to a new starting location after each trial while in an enclosed start box where the animal receives a water reward for running to the end of the track and back into the start box.
Before running the actual task, and after a week of recovery after medial forebrain bundle electrode implantation, mice were placed in a custom box with a nose poke port. They were trained to approach and poke their nose into port, which triggered a brain stimulation reinforcement. “Stim Trigger” controlled the characteristics of the brain stimulation like duration and frequency for each beam break. The other parameters like current and electrode wire combinations were configured on the stimulus isolator. Adjustments to these were made to achieve the highest response rate over a week. The optimal configuration was used as reinforcer for the entire length of the experiment.
To run the task, the researchers used three zone markers. The configuration of the zones was the following: three tracked zones, one to start the trial (i.e., exit from the start box), one rewarded zone and one to mark the arrival at the end of the track. As the actual starting point varied in each trial by design, a manual zone skip feature was employed, triggered by a handheld device specifically designed for advancing presentation slides, to mark the beginning of each trial. The virtual location of the “Start zone” was set so the animal could not reach it, and the trial start was marked by pressing the “Next Cue” button using a handheld presentation clicker. Then, the “Rewarded Zone” became active and the animal could trigger the reward, which in this case was electrical stimulation of the medial forebrain bundle. For this application, the delay feature was used for the “Rewarded Zone” so that the animal must remain in the reward zone for a delay period to or obtained a brain stimulation reward. If the animal advances through the reward zone, or obtained a reward in the zone, then the next nonrewarded zone, called the “End Zone,” becomes active. After reaching the “End Zone,” the animal returns to the start box and consumes a water reward while the track is moved to the next randomly selected start location. Thus, at the end of the session the data for the start of the outward trajectory, entry into and exit from the reward zone, reward timestamp if any, and the time the rodent arrived at the end of the track can be extracted.
A postprocessing script extracts the animal location from the video file and then calculates the velocity for each trial fixed with respect to the position relative to the reward zone. The velocity profile is visually displayed, enabling the user to identify and exclude trials with issues, such as instances of lost tracking that are too extensive to allow for accurate position estimation by the software. This exclusion process can be performed during visualization of the current session. This analysis code is available on request.
The results showed that 3xTg-AD female mice performed significantly worse than non-Tg age-matched controls for the 1.5- and 2-s delays, but not for other delays (Stimmell et al., 2019). The six-month non-Tg female mice were able to identify the location of the reward zone and slowed down in preparation for stopping in the zone, whereas the six-month 3xTg-AD females did not. Overall, the study showed that six-month 3xTg-AD female mice were impaired at spatial reorientation compared with non-Tg mice.
The virtual version of the task is identical except that the animal’s virtual location is restarted automatically to the next starting location from the end of the track (Cushing et al., 2020). During the virtual task, the animal is placed on a tablet surface coated with mineral oil to facilitate smooth movement of its paws. The animal’s head is fixed to ensure that as it walks, the tablet detects the movement of its paws, thereby navigating the animal through the virtual environment, which is projected on the floor and three surrounding wall tablets. Tablet holders are 3D printed with a low-cost resin-based 3D printer and the design files for these tablet holders are available on request. The East, North and West tablets display the rest of the virtual room, which moves along with the animal locomotion. The virtual maze software (Molina et al., 2016) was paired with “Stim Trigger” which controlled the delivery of brain stimulation as described above.
Our next detailed example includes in vivo electrophysiological data. Rats were first trained on an alternation training task, where they learned to shuttle back and forth along a linear track created by walls that connect two opposite zones on the circular platform. Rats receive a brain stimulation that is delivered in at both ends of the linear track. Following the alternation training, rats underwent training on a “random lights task,” where sequences of up to 900 elements were drawn randomly with replacement from the 32 light/reward zones.
Subsequently, rats were trained on the complex spatial sequence task, which involves rats learning to navigate to unmarked locations fixed in space in a specific sequence. Landmarks are distributed around the room for spatial orientation.
We used a sequence (1-2-3-4-1-2-3-5-; Fig. 5A; Movie 1) that had a repeating path segment (1-2-3) followed by one of two distinct actions. Specifically, the rat learned in context 5-1-2-3, to go to 4 for reward, while in context 4-1-2-3 the rat must go to 5.
A, Complex Spatial Sequence Task. Schematic for the complex spatial sequence task. The rat always starts at zone 5 and continues to zones 1-2-3-4-1-2-3-5-. The Zones and Rewarded Zones are set in the “Maze” GUI. As no reward delay was used for this task, as soon as the animal enters the currently active zone, an automatic reward is delivered and the next “Rewarded zone” becomes active following the input order. B, Capture of the behavior during a “cued run.” C, Two plots showing the activity of two hippocampal place cells as a heat plot of occupancy-adjusted firing rate, using an evenly spaced color-map with max rate indicated in red. The firing frequency is also shown. Only areas with a high enough occupancy during the task are represented in the figures.
“Maze” software configuration for Sequence task. At the beginning of the video, a configuration file containing all the necessary settings for Sequence task is loaded. Then, the animal is connected, and electrophysiology recording and live location data are initiated. Subsequently the “Maze” program is initiated and the location of the animal is displayed in the bottom right corner of the software window. In the second part of the video, a rat completes the Sequence task by navigating between the rewarded zones. The path that the animal must travel from memory in order to obtain rewards at each zone is overlaid on the recording for easy visualization.
Thus, each action belongs to two spatial contexts, so navigation to zone 4 or 5 requires a map-to-action transformation. This emulates the common spatial memory problem one encounters when driving through an intersection and remembering the appropriate action given the current route and goals (e.g., turn left to a favorite restaurant vs right to home). Sets of three unguided (“memory”) runs through the complete 1-2-3-4-1-2-3-5 sequence were interleaved with sets of three “cued” runs in which a light at each goal led the rat through the sequence (Fig. 5B). During memory runs, following an error, a light cue directs the rat to the next zone in the sequence. This complex spatial task, while likely engaging the hippocampus, hippocampal activity alone (e.g., splitter cells) is insufficient to predict the rat’s action (Bower et al., 2005). Note, because the sequence makes use of five zones selected from 32 evenly spaced zones distributed around the perimeter of the platform, alternate sequences which match the distance traveled for various elements of the sequence can be created or new sequences can be created by flipping and rotating the sequence to make a novel sequence (we have made use of both of these options in our experiments). The data presented here are the activity of two place cells recorded from the rat’s dorsal hippocampus while it navigated the 8-item sequence on circular platform. The cells have distinct firing patterns focused in a specific location on the platform (Fig. 5C). The animal can be trained by reinforcing their behavior with electrical medial forebrain bundle stimulation or food reward. Rats have been trained under both conditions on the three tasks described in this section. Additionally, this method allowed the researchers to detect the animal’s precise location during behavior, and record the neuronal activity with high temporal resolution, using the recording system.
The next example of the use of our platform is with the object-place paired associate (OPPA) task. (previously described by Sanchez et al., 2019). In brief, rats are trained to travel to the end of a two-arm maze where they have to displace a specific object to obtain a reward out of two possible options. The animal must then navigate back to the center of the maze and go to the other arm. The goal of the current project is to analyze neural activity in the central platform, on each maze arm, and the choice platform where the objects are located. To do this, the Maze software collects accurate timestamps of the location of the animal, as well as entry in zones of interest. The maze software is set-up to detect movements in seven zones, one for each platform at the end of the arms, two for the maze arms, one for the central platform of the maze and two zones outside of the maze to timestamp the manual reward delivery (Fig. 6). To obtain the location of the animal, Bonsai can send location data via UDP, as well as head direction (angle) to the maze software. Each zone entry is sent to the recording system as a TTL to obtain the precise timestamps of the maze events for post processing.
Maze GUI for OPPA task. A, Control of electronic hardware including reward type: none as manual food reward is used. B, Five zones of interest. C, Counters and configuration controls. D, Allocated coordinates for zone location. E, LED test and auto calibration controls. F, Animal ID and comment boxes. G, Real time position tracking with zones (*) in green, unused in black. Unused zones are listed above and below the zones that are positioned on the maze for visualization.
The final example is the map-to-action transformation (MAT) task. Rats first learn to navigate to the reward location (R in Fig. 7, left) from a randomly selected set of seven start locations (S1–S7) using distal cues. The maze is surrounded by landmarks for orientation. Accurate navigation results in food or brain stimulation reward delivered at the reward location. After achieving allocentric condition criterion (over 80% correct trials for at least three out of four sessions), the transformation condition begins. In this condition, the rat is held at each start location in a translucent box for 10 s. This is done for Allocentric and Egocentric conditions to produce parallel datasets for each task. Next, an opaque box is placed over the rat while curtains are pulled obscuring distal cues (Fig. 7, middle). The box is then removed, and the rat must now navigate to R while the distal cues are obscured. The hypothesis is that in this condition the animals must transform the allocentric location of R into an egocentric action sequence to reach the reward.
Map-to-action (MAT) task. Layout for the allocentric (left), transformation (middle), and egocentric conditions (right) of the MAT task. For allocentric and transformation, using “Maze” software, one “Rewarded zone” is used. All the arms S1–S7 and R are marked as “Zones” as well as a center zone. The coordinates as well as the size of the zones can be set in the “Maze” GUI (Fig. 2B,D). No reward delay was used for this task. For egocentric the experimenter determines zone entries manually and triggers the reward using “Send Manually” in “Stim Trigger” (Fig. 3).
After achieving the same criterion for the map-to-action transformation condition, the egocentric paradigm begins where the egocentric relationship between the start location and the rewarded location is maintained (e.g., 4 to 6). However, these locations are randomized for each trial while the distal cues are obscured by the curtains. There are two variations, right turn (Fig. 7), right red and left turn, green. Each rat runs both the right turn and left turn variation separately, and the order is counterbalanced across rats. In this condition, the rat needs to develop a self-centered motor strategy to find the reward and cannot rely on any external cues which are unavailable (distal) or changed every trial (proximal).
Spatial zones are positioned in “Maze” software in order to timestamp zone entry for each of the 8 arms as well as the center of the maze. Some animals are trained with food reward, while others with electrical stimulation of the medial forebrain bundle as the reinforcer. For the animals that receive medial forebrain bundle stimulation reinforcement, the delivery of the brain stimulation pulse train via TTL to the recording system is also timestamped. For all training phases, a nonrewarded zone is positioned in the center of the maze to timestamp entry into the center of the maze. In the case of the allocentric and transformation tasks, the rewarded location remains consistent throughout the session. Two rewarded zones are designated, one located within the correct zone and the other positioned outside of the maze. This ensures that on zone entry, the reward is delivered only once until the experimenter utilizes the “Next Cue” button to reactivate the rewarded zone on the maze. The zone that is positioned off the maze is also used to signal the end of an incorrect trial.
Because of the continuously changing rewarded location for the egocentric task, the “Stim Trigger” software is used to deliver and timestamp the reinforcement delivery for correct trials. Furthermore, manually triggered TTLs can mark events that are not location-dependent, such as the animal placement into boxes. To do this, a wireless relay switch activated by a wireless radio frequency remote control can be used. This relay system is connected to the Neuralynx TTL acquisition board. Ten control rats have successfully completed all three paradigms of the task. Training for each phase takes a similar number of sessions, between 19 and 23.
Discussion
In spatial cognition research, it is essential to have the flexibility to quickly change the experimental setup without having to overhaul the software and hardware. To address this need, we present an inexpensive behavioral platform that can be paired with open-source or commercial software to create a variety of experimental setups. Our platform enables the generation of novel spatial navigation tasks, using positive reinforcement like brain stimulation, liquid, or food reward. We are presently using this platform in our laboratory to collect data using eight different mazes/experimental paradigms and there are countless more experiments that can be designed and run using this platform.
Our software includes two highly flexible programs. The Maze program can use up to 32 user defined zones and deliver rewards automatically on zone entry or track progress through the zones. This program is highly synergistic with electrophysiology acquisition systems like Neuralynx and Open Ephys, as it timestamps zone entries and other behaviors using TTL signals triggered by the Maze program and delivered to Digital Lynx SX or Open Ephys Acquisition Board. The second program, “Stim Trigger,” can control brain stimulation using any type of equipment that can be paired with an Arduino board, such as a nose-poke and a virtual reality maze.
Our platform distinguishes itself from other commercial and open-source software options for animal behavioral tracking by offering more than just tracking capabilities. For example, EthoVision is a commonly used video tracking software, capturing real-time movements and yielding detailed behavioral data. There are available extensions that carry a higher cost, such as EthoVisionXT, that incorporate a “Trial and hardware control” module with a USB-IO box. Noldus offers an extensive catalog for behavioral rodent studies; however, its modular software’s limitations arise when crucial modules are inaccessible, and its proprietary nature hampers cross-platform compatibility.
Another example is ANY-maze, a versatile tool that supports varied animal movement tracking and behavior analyses, even within operant conditioning chambers. ANY-maze’s complete license includes updates and add-ons, and interface devices are available for hardware control. Although versatile, ANY-maze is not free and lacks intrinsic ability to synchronize with other extensively used software platforms like Python or MATLAB.
Finally, our platform seamlessly integrates with Bonsai-RX and DeepLabCut, enhancing its capabilities even further. Bonsai-RX excels in real-time behavioral analysis and closed-loop experiments. It facilitates complex experimental designs through a user-friendly interface and pre-built components. Its adaptability stems from a modular structure that integrates hardware and software components, making it our choice for implementation. It allows researchers to perform real-time behavioral analysis and closed-loop experiments, while DeepLabCut’s deep learning algorithms enable accurate and automated tracking of key points on the subject’s body, allowing for precise analysis of movement and behavior. The synergy between this platform and these software solutions employing UDP connections facilitates the implementation of complex experimental setups, incorporating multiple data streams such as video tracking, electrophysiology, and optogenetics, offering researchers extensive possibilities for data analysis and visualization.
We hope that the ingenuity of this approach goes beyond its impact on spatial cognition research. The adaptable behavioral platform introduced in this paper offers a wealth of possibilities for diverse nonspatial testing scenarios. By leveraging the “Maze” and “Stim Trigger” software components, researchers can seamlessly extend their investigations to many domains. For instance, in the realm of cognitive psychology, this platform could facilitate studies related to attention, memory, and decision-making. The capacity for precise time stamping of neural events coupled with automated reward delivery could also be harnessed to study reinforcement learning, habit formation, and associative learning processes. Moreover, within the field of behavioral neuroscience, the platform’s flexibility enables exploration of sensory perception or motor control. Researchers could use the brain stimulation control capability to investigate neural circuitry underlying other complex behaviors. In the context of pharmacological research, the platform might serve as a valuable tool for assessing drug effects on behavior, thereby aiding in the development of new treatments for a wide range of disorders. Classic self-stimulation paradigms can also be easily and inexpensively conducted using “Stim Trigger,” which opens many possibilities in the addiction field.
The purpose of sharing these programs is 2-fold. First, to make the software available in its current state, and second, to solicit feedback from the broader community beyond our own groups, with the aim of enhancing and expanding our software package, including into fields beyond spatial cognition. We have focused on making our software compatible with open-source software and hardware like Bonsai and Open Ephys. A potential area for expansion is with another cutting-edge open-source technique, the UCLA Miniscope (Ghosh et al., 2011). It can be paired with our “Maze” software using the Miniscope V4-Data Acquisition System to synchronize Neuralynx or Bonsai (with a blinking LED for synchronization) video acquisition with the physiological data and timestamp location information. We seek to continue developing these tools and collaborating with other researchers to increase the possible applications of our work.
Limitations
Our software has limitations when it comes to compatibility with all experimental designs. For instance, the current version only allows for the delivery of rewards on entry to a zone or zones sequentially, but two rewarded zones cannot be available simultaneously, as we have not yet encountered a need for this functionality.
While our software has been effectively used in both rats and mice, including several publications involving mice, it is important to note that the physical maze platform described in the methods section was primarily designed with larger rodents, such as rats, in mind. To address this, we propose the creation of scaled-down versions of the platform specifically tailored for smaller rodents. These scaled-down versions would function in a similar manner to the original platform, with the location and size of the zones adjusted proportionally based on the size of the camera frame. No modification of the software is needed to make use of a scaled down platform.
We also plan to address compatibility issues. While the software and hardware are compatible with open-source software and hardware such as Bonsai and Open Ephys, we recognize that MATLAB is not freely available. Therefore, we aim to transition to other programming languages such as Python in the future. In the meantime, MATLAB can be used to compile the program (and we have made available a compiled version of the current program), so MATLAB is only needed when making edits to the program. While we acknowledge that a single platform may not be able to cover all experimental designs, the software’s flexible structure enables us and other users to create custom versions quickly.
Footnotes
The authors declare no competing financial interests.
This work was supported by NIH Grants R01 AG070094, F32 MH099682, K99/R00 AG049090, and FL DOH 20A09 (to A.A.W.); MH46823-16 (to B.L.M.); R01 AA029700 (to B.J.C.); and F31 AA030711 (to L.M.S.).
This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.