August 8, 2018
In HardwareX, an open access journal for designing, building and customizing opensource scientific hardware, Martin A. Raymond and colleagues share their design for a user-constructed, low-cost lickometer.
Researchers interested in ingestive behaviors of rodents commonly use licking behavior as a readout for the amount of fluid a subject consumes, as recorded by a lickometer. Commercially available lickometers are powerful tools to measure this behavior, but can be expensive and often require further customization. The authors offer their own design for an opensource lickometer that utilizes readily available or customizable components such as a PC sound card and 3D printed drinking bottle holder. The data from this device is collected by Audacity, and opensource audio program, which is then converted to a .csv format which can be analyzed using an R script made available by the authors to assess various features of licking microstructure. A full bill of materials, instructions for assembly and links to design files are available in the paper.
Check out the full publication here!
Raymond, M. A., Mast, T. G., & Breza, J. M. (2018). An open-source lickometer and microstructure analysis program. HardwareX, 4. doi:10.1016/j.ohx.2018.e00035
July 23, 2018
OpenBehavior has been covering open-source neuroscience projects for a few years, and we are always thrilled to see projects that are well documented and can be easily reproduced by others. To further this goal, we have formed a collaboration with Hackaday.io, who have provided a home for OpenBehavior on their site. This can be found at: https://hackaday.io/OpenBehavior, where we currently have 36 projects listed ranging from electrophysiology to robotics to behavior. We are excited about this collaboration because it provides a straightforward way for people to document their projects with instructions, videos, images, data, etc. Check it out, see what’s there, and if you want your project linked to the OpenBehavior page simply tag it as “OPENBEHAVIOR” or drop us a line at the Hackaday page.
Note: This collaboration between OpenBehavior and Hackaday.io is completely non-commercial, meaning that we don’t pay Hackaday.io for anything, nor do we receive any payments from them. It’s simply a way to further our goal of promoting open-source neuroscience tools and their goal of growing their science and engineering community.
July 16, 2018
In a special issue of Journal of Neural Engineering, Dominique Martinez and colleagues their share design for NeRD, an open source neural recording device for wireless transmission of local field potential (LFP) data in in freely-behaving animals.
Electrophysiological recording of local field potentials in freely-behaving animals is a prominent tool used by researchers for assessing the neural basis of behavior. When performing these recordings, cables are commonly used to transmit data to the recording equipment, which tethers the animals and can interfere with natural behavior. Wireless transmission of LFP data has the advantage of removing the cable between the animal and the recording equipment, but is hampered by the large number of data to be transmitted at a relatively high rate.
To reduce transmission bandwidth, Martinez et al. propose an encoder/decoder algorithm based on adaptive non-uniform quantization. As proof-of- concept, they developed a NeRD prototype that digitally transmits eight channels encoded at 10 kHz with 2 bits per sample. This lightweight device occupies a small volume and is powered with a small battery allowing for 2h 40min of autonomy. The power dissipation is 59.4 mW for a communication range of 8 m and transmission losses below 0.1%. The small weight and low power consumption offer the possibility of mounting the entire device on the head of a rodent without resorting to a separate head-stage and battery backpack. The use of adaptive quantization in the wireless transmitting neural implant allows for lower transmission bandwidths, preservation of high signal fidelity, and preservation of fundamental frequencies in LFPs from a compact and lightweight device.
Martinez, D., Clément, M., Messaoudi, B., Gervasoni, D., Litaudon, P., & Buonviso, N. (2018). Adaptive quantization of local field potentials for wireless implants in freely moving animals: An open-source neural recording device. Journal of Neural Engineering, 15(2), 025001. doi:10.1088/1741-2552/aaa041
June 25, 2018
Andreas Genewsky and colleagues from the Max Planck Institute of Psychiatry have shared the design, construction and validation of a simplified, low-cost, radar-based motion detector for home cage activity monitoring in mice. This simple, open-source device allows for motion detection without visual contact to the animal and can be used with various cage types. It features a custom printed circuit board and motion detector shield for Arduino, which saves raw activity and timestamped data in CSV files onto an SD card; the authors also provide a Python script for data analysis and generation of actograms. This device offers a cost-effective, DIY alternative to optical imaging of home-cage activity.
Read more from the Journal of Biomedical Engineering publication!
Genewsky, A., Heinz, D. E., Kaplick, P. M., Kilonzo, K., & Wotjak, C. T. (2017). A simplified microwave-based motion detector for home cage activity monitoring in mice. Journal of Biological Engineering,11(1). doi:10.1186/s13036-017-0079-y
June 15, 2018
In a recent preprint on BioRxiv, Alessio Buccino and colleagues from the University of Oslo provide a step-by-step guide for setting up an open source, low cost, and adaptable system for combined behavioral tracking, electrophysiology, and closed-loop stimulation. Their setup integrates Bonsai and Open Ephys with multiple modules they have developed for robust real-time tracking and behavior-based closed-loop stimulation. In the preprint, they describe using the system to record place cell activity in the hippocampus and medial entorhinal cortex, and present a case where they used the system for closed-loop optogenetic stimulation of grid cells in the entorhinal cortex as examples of what the system is capable of. Expanding the Open Ephys system to include animal tracking and behavior-based closed-loop stimulation extends the availability of high-quality, low-cost experimental setup within standardized data formats.
Read more on BioRxiv, or on GitHub!
Buccino A, Lepperød M, Dragly S, Häfliger P, Fyhn M, Hafting T (2018). Open Source Modules for Tracking Animal Behavior and Closed-loop Stimulation Based on Open Ephys and Bonsai. BioRxiv. http://dx.doi.org/10.1101/340141
June 12, 2018
In a recent publication in the Frontiers in Systems Neuroscience, Solari and colleagues of the Hungarian Academy of Sciences and Semmelweis University have shared the following about a behavioral setup for temporally controlled rodent behavior. This arrangement allows for training of head-fixed animals with calibrated sound stimuli, precisely timed fluid and air puff presentations as reinforcers. It combines microcontroller-based behavior control with a sound delivery system for acoustic stimuli, fast solenoid valves for reinforcement delivery and a custom-built sound attenuated chamber, and is shown to be suitable for combined behavior, electrophysiology and optogenetics experiments. This system utilizes an optimal open source setup of both hardware and software through using Bonsai, Bpod and OpenEphys.
Read more here!
Solari N, Sviatkó K, Laszlovszky T, Hegedüs P and Hangya B (2018). Open Source Tools for Temporally Controlled Rodent Behavior Suitable for Electrophysiology and Optogenetic Manipulations. Front. Syst. Neurosci. 12:18. doi: 10.3389/fnsys.2018.00018
June 8, 2018
OpenBehavior has shared a variety of popular open-source tracking software, and there’s another to add to the list: ToxTrac!
Alvaro Rodriguez and colleagues from Umeå University in Umeå, Sweden, have developed ToxTrac, an open-source Windows program optimized for high-speed tracking of animals. It uses an advanced tracking algorithm that requires no specific knowledge of the geometry of tracked bodies and can therefore be used for a variety of species. ToxTrac can also track multiple bodies in multiple arenas simultaneously, while maintaining individual identification. The software is fast, operating at a rate >25 frames per second, and robust against false positives. ToxTrac generates useful statistics and heat maps in real scale that can be exported in image, text and excel formats to provide useful information about locomotor activity in rodents, insects, fish, etc.
Learn more about ToxTrac here: https://doi.org/10.1111/2041-210X.12874
Or Download ToxTrac software here: https://toxtrac.sourceforge.io
Rodriguez A, Zhang H, Klaminder J, Brodin T, Andersson PL, Andersson M. ToxTrac: A fast and robust software for tracking organisms. Methods Ecol Evol. 2018;9:460–464. https://doi.org/10.1111/2041-210X.12874
June 6, 2018
This post is relevant for MedPC users who also use MatLab or Python for data analysis.
We recently became aware that many MedPC users are not saving precise times for behavioral events from their experiments. A method called time-event codes was worked out around 2000 by Russ Church and his group at Brown, working with MedAssociates. Marcelo Caetano, a former postdoc in the Laubach Lab at Yale, incorporated this approach into an existing MatLab function (MedParse, written by Kumar Narayanan during his PhD training in the Laubach Lab at Yale). More recently, the code was ported to Python by Kyra Swanson, a Phd student in the Laubach Lab at American University. It is available at https://github.com/LaubachLab/MedParse. MedPC code for saving precise times of behavioral events (example in MedPC Template) and MatLab and Python functions are provided that convert MedPC data (see the template) into “time-event codes,” i.e., a two-column matrix with times (column 1) and events (column 2).
April 2, 2018
Check out the Ethoscopes platform!
Ethoscopes enable high-throughput analysis of behavior in Drosophila and other animals for <$100. The system is capable of real-time video tracking, is based on raspberry pi, and even has its own R package for data analysis. All software and build specifications are available at http://lab.gilest.ro/ethoscope.
March 8, 2018
Robyn A. Grant, from Manchester Metropolitan University, has shared the following on Twitter regarding the development of the LocoWhisk arena:
“Come help me develop my new arena. Happy to hear from anyone looking to test it or help me develop it further.”
The LocoWhisk system is a new, portable behavioural set-up that incorporates both gait analysis (using a pedobarograph) and whisker movements (using high-speed video camera and infrared light source). The system has so far been successfully piloted on many rodent models, and would benefit from further validation and commercialisation opportunities.
Learn more here: https://crackit.org.uk/locowhisk-quantifying-rodent-exploration-and-locomotion-behaviours