Cargando…

Real-Time Closed-Loop Feedback in Behavioral Time Scales Using DeepLabCut

Computer vision approaches have made significant inroads into offline tracking of behavior and estimating animal poses. In particular, because of their versatility, deep-learning approaches have been gaining attention in behavioral tracking without any markers. Here, we developed an approach using D...

Descripción completa

Detalles Bibliográficos
Autores principales: Sehara, Keisuke, Zimmer-Harwood, Paul, Larkum, Matthew E., Sachdev, Robert N. S.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Society for Neuroscience 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8174057/
https://www.ncbi.nlm.nih.gov/pubmed/33547045
http://dx.doi.org/10.1523/ENEURO.0415-20.2021
Descripción
Sumario:Computer vision approaches have made significant inroads into offline tracking of behavior and estimating animal poses. In particular, because of their versatility, deep-learning approaches have been gaining attention in behavioral tracking without any markers. Here, we developed an approach using DeepLabCut for real-time estimation of movement. We trained a deep-neural network (DNN) offline with high-speed video data of a mouse whisking, then transferred the trained network to work with the same mouse, whisking in real-time. With this approach, we tracked the tips of three whiskers in an arc and converted positions into a TTL output within behavioral time scales, i.e., 10.5 ms. With this approach, it is possible to trigger output based on movement of individual whiskers, or on the distance between adjacent whiskers. Flexible closed-loop systems like the one we have deployed here can complement optogenetic approaches and can be used to directly manipulate the relationship between movement and neural activity.