June 15, 2018
In a recent preprint on BioRxiv, Alessio Buccino and colleagues from the University of Oslo provide a step-by-step guide for setting up an open source, low cost, and adaptable system for combined behavioral tracking, electrophysiology, and closed-loop stimulation. Their setup integrates Bonsai and Open Ephys with multiple modules they have developed for robust real-time tracking and behavior-based closed-loop stimulation. In the preprint, they describe using the system to record place cell activity in the hippocampus and medial entorhinal cortex, and present a case where they used the system for closed-loop optogenetic stimulation of grid cells in the entorhinal cortex as examples of what the system is capable of. Expanding the Open Ephys system to include animal tracking and behavior-based closed-loop stimulation extends the availability of high-quality, low-cost experimental setup within standardized data formats.
Read more on BioRxiv, or on GitHub!
Buccino A, Lepperød M, Dragly S, Häfliger P, Fyhn M, Hafting T (2018). Open Source Modules for Tracking Animal Behavior and Closed-loop Stimulation Based on Open Ephys and Bonsai. BioRxiv. http://dx.doi.org/10.1101/340141
June 12, 2018
In a recent publication in the Frontiers in Systems Neuroscience, Solari and colleagues of the Hungarian Academy of Sciences and Semmelweis University have shared the following about a behavioral setup for temporally controlled rodent behavior. This arrangement allows for training of head-fixed animals with calibrated sound stimuli, precisely timed fluid and air puff presentations as reinforcers. It combines microcontroller-based behavior control with a sound delivery system for acoustic stimuli, fast solenoid valves for reinforcement delivery and a custom-built sound attenuated chamber, and is shown to be suitable for combined behavior, electrophysiology and optogenetics experiments. This system utilizes an optimal open source setup of both hardware and software through using Bonsai, Bpod and OpenEphys.
Read more here!
Solari N, Sviatkó K, Laszlovszky T, Hegedüs P and Hangya B (2018). Open Source Tools for Temporally Controlled Rodent Behavior Suitable for Electrophysiology and Optogenetic Manipulations. Front. Syst. Neurosci. 12:18. doi: 10.3389/fnsys.2018.00018
May 21, 2018
Meaghan Creed has developed a novel device for assessing preferences by mice among fluids in their homecages, i.e. two-bottle choice test. She shared the design on http://hackaday.io and contributed the summary of it below.
Often in behavioral neuroscience, we need to measure how often and how much a mouse will consume multiple liquids in their home cage. Examples include sucrose preference tasks in models of depression, or oral drug self-administration (ie. Morphine, opiates) in the context of addiction. Classically, two bottles are filled with liquids and volumes are manually recorded at a single time point. Here, we present a low-cost, two-sipper apparatus that mounts on the inside of a standard mouse cage. Interactions are detected using photointerrupters at the base of each sipper which are logged to an SD card using a standard Arduino. Sippers are constructed from 15 mL conical tubes which allows additional volumetric measurements, the rest of the holding apparatus is 3D printed, and the apparatus is constructed with parts from Arduino and Sparkfun. This automated approach allows for high temporal resolution collected over 24 hours, allowing measurements of patterns of intake in addition to volume measurements. Since we don’t need to manually weigh bottles we can do high-throughput studies, letting us run much larger cohorts.
This is designed such that each set of 2 sippers uses its own Arduino and SD card. With a bit of modification to the code one Arduino Uno can be programmed to log from 6 cages onto the same SD card. Arduino compatible boards with more GPIOs (like Arduino Mega) can log from up to 56 sippers on one Arduino.
March 9, 2018
O’Leary and colleagues describe an open-source touch-screen for rodent behavioral testing. The manuscript is well documented and includes all of the parts needed to build the system on your own. Very useful methods for testing cognitive function and relating findings across species (rodents, primates, humans). Congrats to the authors on setting a high standard for open-source neuroscience!
O’Leary, J.D., O’Leary, O.F., Cryan, J.F. et al. Behav Res (2018). https://doi.org/10.3758/s13428-018-1030-y
March 1, 2018
From the Kravitz lab at the NIH comes a simple device for dispensing pre-measured quantities of food at regular intervals throughout the day. Affectionately known as “SnackClock”, this device uses a 24-hour clock movement to rotate a dispenser wheel one revolution per day. The wheel contains 12 compartments, which allows the device to dispense 12 pre-measured “snacks” at regular 2 hour intervals. The Kravitz lab has used this device to dispense high-fat diet throughout the day, rather than giving mice one big piece once per day. The device is very simple to build and use, requiring just two 3D printed parts and a ~$10 clock movement. There is no microcontroller or coding required for this device, and it runs on one AA battery for >1 year. The 3D files are supplied and can be edited to fit SnackClock in different brands of caging, or to adjust the number of snack compartments. With additional effort the clock movement could be replaced by a stepper motor to allow for dispensing at irregular or less frequent intervals.
An interesting summary of recent methods for monitoring behavior in rodents was published this week in Nature.The article mentions Lex Kravitz and his lab’s efforts on the Feeding Experimentation Device (FED) and also OpenBehavior. Check it out: https://www.nature.com/articles/d41586-018-02403-5
February 6, 2018
Brian Isett, who is now at Carnegie Mellon, has kindly shared the following tutorial regarding the creation and implementation of a Rodent Running Disk he designed while at University of California, Berkeley.
“Awake, naturalistic behavior is the gold standard for many neuroscience experiments. Increasingly, researchers using the mouse model system strive to achieve this standard while also having more control than a freely moving animal. Using head-fixation, a mouse can be positioned very precisely relative to ongoing stimuli, but often at the cost of naturalism. One way to overcome this problem is to use the natural running of the mouse to control stimulus presentation in a closed-loop “virtual navigation” environment. This combination allows for awake, naturalistic behavior, with the added control of head-fixation. A key element of this paradigm is to have a very fast way of decoding mouse locomotion.
In this tutorial, we describe using an acrylic disk mounted to an optical encoder to achieve fast locomotion decoding. Using an Arduino to decode the TTL pulses coming from the optical encoder, real-time, closed-loop stimuli can be easily presented to a head-fixed mouse. This ultimately allowed us to present tactile gratings to a mouse performing a whisker-mediated texture discrimination task as a “virtual foraging task” — tactile stimuli moved past the whiskers synchronously with mouse locomotion. But the design is equally useful for measuring mouse running position and speed in a very precise way.”
The tutorial may be found here.
Isett, B.R., Feasel, S.H., Lane, M.A., and Feldman, D.E. (2018). Slip-Based Coding of Local Shape and Texture in Mouse S1. Neuron 97, 418–433.e5.
January 8th, 2018
The de Bivort lab and FlySorter, LLC are happy to share on OpenBehavior their open-source Drosophila handling platform, called MAPLE: Modular Automated Platform for Large-Scale Experiments.
Drosophila Melanogaster has proven a valuable genetic model organism due to the species’ rapid reproduction, low-maintenance, and extensive genetic documentation. However, the tedious chore of handling and manually phenotyping remains a limitation with regards to data collection. MAPLE: a Modular Automated Platform for Large-Scale Experiments provides a solution to this limitation.
MAPLE is a Drosophila-handing robot that boasts a modular design, allowing the platform to both automate diverse phenotyping assays and aid with lab chores (e.g., collecting virgin female flies). MAPLE permits a small-part manipulator, a USB digital camera, and a fly manipulator to work simultaneously over a platform of flies. Failsafe mechanisms allow users to leave MAPLE unattended without risking damage to MAPLE or the modules.
The physical platform integrates phenotyping and animal husbandry to allow end-to-end experimental protocols. MAPLE features a large, physically-open workspace for user convenience. The sides, top, and bottom are made of clear acrylic to allow optical phenotyping at all time points other than when the end-effector carriages are above the modules. Finally, the low cost and scalability allow large-scale experiments ($3500 vs hundreds of thousands for a “fly-flipping” robot).
MAPLE’s utility and versatility were demonstrated through the execution of two tasks: collection of virgin female flies, and a large-scale longitudinal measurement of fly social networks and behavior.
Links to materials:
Raw data and analysis scripts
De Bivort Lab Site
January 3rd, 2018
The following behavioral platform was developed and published by Xinfeng Chen and Haohong Li, from Huazhong University of Science and Technology, Wuhan, China
ArControl: Arduino Control Platform is a comprehensive behavioral platform developed to deliver stimuli and monitor responses. This easy-to-use, high-performance system uses an Arduino UNO board and a simple drive circuit along with a stand-along GUI application. Experimental data is automatically recorded with the built-in data acquisition function and the entire behavioral schedule is stored within the Arduino chip. Collectively, this makes ArControl a “genuine, real-time system with high temporal resolution”. Chen and Li have tested ArControl using a Go/No-Go task and a probabilistic switching behavior task. The results of their work show that ArControl is a reliable system for behavioral research.
Source codes and PCB drafts may be found here: ArControl Github
November 28, 2017
Airtrack was developed in LARKUM Lab by Mostafa Nashaat, Hatem Oraby, Robert Sachdev, York Winter and Matthew Larkum. Alexander Schill, engineer at Charité workshop (CWW) had a significant contribution to the design of the platform and the airtrack table.
Airtrack is a head-fixed behavioral environment that uses a lightweight physical maze floating on an air table that moves around the animal’s body under the direct control of the animal itself, solving many problems associated with using virtual reality for head-fixed animals.
More Information can be found at http://www.neuro-airtrack.com/
Nashaat, MA, Oraby, H, Sachdev, RNS, Winter, Y, Larkum, ME. (2016).
Air-Track: a real-world floating environment for active sensing in head-fixed mice.
Journal of Neurophysiology 116 (4) 1542-1553; DOI:10.1152/jn.00088.2016