Note: AIM is informally referred to as DNAMIC (Dartmouth Nautiyal Arduino Modular Instrumental Chamber) in the lab. Therefore I might use the two terms (AIM / DNAMIC) interchangeably to refer to the same thing.
Figure 1: Mouse interacting with the AIM box - Currently retrieving water reward from the middle port.
Figure 2: Diagonal View of AIM
Figure 5: AIM boxes in operation within our lab satellite room. Data is collected automatically by Processing software to a computer in the next room.
AIM is a customizable, low-cost operant box to train and test various paradigms in mice. Made with Arduino microprocessor, 3D printed parts, and various off-the-shelf components. Enables 24-hour data collection, significantly reducing total experiment duration from months to a couple weeks, thus facilitating controlled studies on adolescent mice and revealing high temporal resolution data on mice behavior.
Most importantly, this is an open-source project - any contributions / suggestions / feedback are welcome and appreciated!
AIM has a couple of key benefits for any behavioral neuroscience labs trying to study adolescene in rodents.
-
1. Decreased Experiment Time
- Because subjects have 24 hour access to a specific paradigm, they learn quickly. Paradigms that typically takes months to train can be completed in just a matter of two weeks.
-
2. Inexpensive compared to traditional Med Associates Boxes.
- All the materials used to make one AIM box only cost about $200. The software used to run the hardware is also free and open-source. In contrast, a typical Med Associates Box cost around $4000. The software rig that can control up to 8 Med Associates Boxes costs about $6,800.
There are other key benefits to using AIM and you can find a complete list of benefits and advantages in the Introduction. But most importantly, it saves YOU time and money, allowing you to focus on other important lab tasks and allocate grant money to other experiments.
The purpose of this repository is to provide hardware manuals and documentations for anyone who would like to build their own AIM for use in behavioral neuroscience research. This manual is intended for behavioral neuroscientists without previous experience in engineering or computer science. Therefore the manuals sometimes include explanations on the inner workings of major components used in AIM.
AIM is divided into three different repositories. This repository contains the hardware build instructions for AIM. The second repository contains the corresponding behavioral paradigms for AIM. The last repository contains Python codes for data analysis of the behavioral output. Data analysis code is currently in development to create a python package.
The below chapters are rendered via the Jupyter nbviewer and is read-only.
-
Chapter 0 - Background and Introduction Why we need Arduino Instrumental Modules (AIM)
- Feeding Experimentation Device (FED) - Kravitz Lab
- CombiCage - Loos Lab
- ArControl - Li Lab
- Openmaze.org - open source hardware and software for behavioral neuroscience
- Openbehavior.com - repository of open source tools for advancing behavioral neuroscience research]
-
Inaccurate / incorrect information regarding components or explanations
-
Typos or grammar mistakes
-
Component suggestions
-
Your own tips, tricks, and experiences for any hardware projects.
-
If you decide to make AIMs yourself, send us photos and any hacks/design modifications that you made! We would love to see how this project expands!
Feel free to open an issue or submit a pull request. All contributions are welcome! - there is no minor contribution. If you're unfamiliar with pull request workflow, feel free to email me with your contributions to jun.ho.lee@dartmouth.edu.
A big thanks to Andrew R. Alvarenga for his generous support and suggestions regarding hardware. Thank you for also allowing me to use the machine shop and test out different prototypes of the 3D printed parts on the 3D printer.
Thank you to undergraduate researchers Bonnie Shea and Selin Capan for assisting with daily running of the code.
For any questions or issues, contact the main author, Jun Ho Lee at jun.ho.lee@dartmouth.edu or the corresponding author Katherine Nautiyal at katherine.nautiyal@dartmouth.edu.