Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies

Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of neuroscience methods 2009-06, Vol.180 (2), p.224-233
Hauptverfasser: Peikon, Ian D., Fitzsimmons, Nathan A., Lebedev, Mikhail A., Nicolelis, Miguel A.L.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 233
container_issue 2
container_start_page 224
container_title Journal of neuroscience methods
container_volume 180
creator Peikon, Ian D.
Fitzsimmons, Nathan A.
Lebedev, Mikhail A.
Nicolelis, Miguel A.L.
description Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100 fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion.
doi_str_mv 10.1016/j.jneumeth.2009.03.010
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_67275955</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0165027009001605</els_id><sourcerecordid>20079831</sourcerecordid><originalsourceid>FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</originalsourceid><addsrcrecordid>eNqFkcFu1DAQhi1ERZfCK1Q-cWrCOHac9Q1U0YJUiUuRuFmOPWG9xE6xnUq98Q68IU9SV7uIY08j_fP9M5r5CTln0DJg8v2-3UdcA5Zd2wGoFngLDF6QDdsOXSOH7feXZFPBvoFugFPyOuc9AAgF8hU5ZUpI0TOxIel2lxAb5wPG7Jdo5gtq1rIEU9Bd0IRmbkpt0nvvcKH5IRcMdFoSLcnYnz7-oLMPIw1LqW7qIx2T8fHv7z_B2J2PWKWCaTIWaS6r85jfkJPJzBnfHusZ-Xb16fbyc3Pz9frL5cebxnI1lMYZLkANCKMAxiWK0Y3jhF03TY4LLl0_WgnKwWAFYD2s722nqu46y0cm-Bl5d5h7l5ZfK-aig88W59lEXNas5dANver7Z8H64EFtOaugPIA2LTknnPRd8sGkB81AP8Wi9_pfLE8upYHrGks1nh83rGNA9992zKECHw4A1ofce0w6W4_RovMJbdFu8c_teASojqTs</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>20079831</pqid></control><display><type>article</type><title>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</title><source>MEDLINE</source><source>Elsevier ScienceDirect Journals</source><creator>Peikon, Ian D. ; Fitzsimmons, Nathan A. ; Lebedev, Mikhail A. ; Nicolelis, Miguel A.L.</creator><creatorcontrib>Peikon, Ian D. ; Fitzsimmons, Nathan A. ; Lebedev, Mikhail A. ; Nicolelis, Miguel A.L.</creatorcontrib><description>Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100 fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion.</description><identifier>ISSN: 0165-0270</identifier><identifier>EISSN: 1872-678X</identifier><identifier>DOI: 10.1016/j.jneumeth.2009.03.010</identifier><identifier>PMID: 19464514</identifier><language>eng</language><publisher>Netherlands: Elsevier B.V</publisher><subject>Action Potentials - physiology ; Algorithms ; Animal movement ; Animal posture ; Animals ; Biomechanical Phenomena - physiology ; Brain - physiology ; Brain–machine interface ; Computer Simulation ; Extremities - physiology ; Image Processing, Computer-Assisted - instrumentation ; Image Processing, Computer-Assisted - methods ; Macaca mulatta ; Motor Cortex - physiology ; Movement - physiology ; Neurons - physiology ; Neurophysiology - instrumentation ; Neurophysiology - methods ; Pattern Recognition, Automated - methods ; Primates ; Range of Motion, Articular - physiology ; Real-time systems ; Signal Processing, Computer-Assisted ; Time Factors ; User-Computer Interface ; Video Recording - instrumentation ; Video Recording - methods ; Video tracking</subject><ispartof>Journal of neuroscience methods, 2009-06, Vol.180 (2), p.224-233</ispartof><rights>2009 Elsevier B.V.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</citedby><cites>FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S0165027009001605$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65306</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/19464514$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Peikon, Ian D.</creatorcontrib><creatorcontrib>Fitzsimmons, Nathan A.</creatorcontrib><creatorcontrib>Lebedev, Mikhail A.</creatorcontrib><creatorcontrib>Nicolelis, Miguel A.L.</creatorcontrib><title>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</title><title>Journal of neuroscience methods</title><addtitle>J Neurosci Methods</addtitle><description>Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100 fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion.</description><subject>Action Potentials - physiology</subject><subject>Algorithms</subject><subject>Animal movement</subject><subject>Animal posture</subject><subject>Animals</subject><subject>Biomechanical Phenomena - physiology</subject><subject>Brain - physiology</subject><subject>Brain–machine interface</subject><subject>Computer Simulation</subject><subject>Extremities - physiology</subject><subject>Image Processing, Computer-Assisted - instrumentation</subject><subject>Image Processing, Computer-Assisted - methods</subject><subject>Macaca mulatta</subject><subject>Motor Cortex - physiology</subject><subject>Movement - physiology</subject><subject>Neurons - physiology</subject><subject>Neurophysiology - instrumentation</subject><subject>Neurophysiology - methods</subject><subject>Pattern Recognition, Automated - methods</subject><subject>Primates</subject><subject>Range of Motion, Articular - physiology</subject><subject>Real-time systems</subject><subject>Signal Processing, Computer-Assisted</subject><subject>Time Factors</subject><subject>User-Computer Interface</subject><subject>Video Recording - instrumentation</subject><subject>Video Recording - methods</subject><subject>Video tracking</subject><issn>0165-0270</issn><issn>1872-678X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2009</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNqFkcFu1DAQhi1ERZfCK1Q-cWrCOHac9Q1U0YJUiUuRuFmOPWG9xE6xnUq98Q68IU9SV7uIY08j_fP9M5r5CTln0DJg8v2-3UdcA5Zd2wGoFngLDF6QDdsOXSOH7feXZFPBvoFugFPyOuc9AAgF8hU5ZUpI0TOxIel2lxAb5wPG7Jdo5gtq1rIEU9Bd0IRmbkpt0nvvcKH5IRcMdFoSLcnYnz7-oLMPIw1LqW7qIx2T8fHv7z_B2J2PWKWCaTIWaS6r85jfkJPJzBnfHusZ-Xb16fbyc3Pz9frL5cebxnI1lMYZLkANCKMAxiWK0Y3jhF03TY4LLl0_WgnKwWAFYD2s722nqu46y0cm-Bl5d5h7l5ZfK-aig88W59lEXNas5dANver7Z8H64EFtOaugPIA2LTknnPRd8sGkB81AP8Wi9_pfLE8upYHrGks1nh83rGNA9992zKECHw4A1ofce0w6W4_RovMJbdFu8c_teASojqTs</recordid><startdate>20090615</startdate><enddate>20090615</enddate><creator>Peikon, Ian D.</creator><creator>Fitzsimmons, Nathan A.</creator><creator>Lebedev, Mikhail A.</creator><creator>Nicolelis, Miguel A.L.</creator><general>Elsevier B.V</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7TK</scope><scope>7X8</scope></search><sort><creationdate>20090615</creationdate><title>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</title><author>Peikon, Ian D. ; Fitzsimmons, Nathan A. ; Lebedev, Mikhail A. ; Nicolelis, Miguel A.L.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2009</creationdate><topic>Action Potentials - physiology</topic><topic>Algorithms</topic><topic>Animal movement</topic><topic>Animal posture</topic><topic>Animals</topic><topic>Biomechanical Phenomena - physiology</topic><topic>Brain - physiology</topic><topic>Brain–machine interface</topic><topic>Computer Simulation</topic><topic>Extremities - physiology</topic><topic>Image Processing, Computer-Assisted - instrumentation</topic><topic>Image Processing, Computer-Assisted - methods</topic><topic>Macaca mulatta</topic><topic>Motor Cortex - physiology</topic><topic>Movement - physiology</topic><topic>Neurons - physiology</topic><topic>Neurophysiology - instrumentation</topic><topic>Neurophysiology - methods</topic><topic>Pattern Recognition, Automated - methods</topic><topic>Primates</topic><topic>Range of Motion, Articular - physiology</topic><topic>Real-time systems</topic><topic>Signal Processing, Computer-Assisted</topic><topic>Time Factors</topic><topic>User-Computer Interface</topic><topic>Video Recording - instrumentation</topic><topic>Video Recording - methods</topic><topic>Video tracking</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Peikon, Ian D.</creatorcontrib><creatorcontrib>Fitzsimmons, Nathan A.</creatorcontrib><creatorcontrib>Lebedev, Mikhail A.</creatorcontrib><creatorcontrib>Nicolelis, Miguel A.L.</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Neurosciences Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>Journal of neuroscience methods</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Peikon, Ian D.</au><au>Fitzsimmons, Nathan A.</au><au>Lebedev, Mikhail A.</au><au>Nicolelis, Miguel A.L.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</atitle><jtitle>Journal of neuroscience methods</jtitle><addtitle>J Neurosci Methods</addtitle><date>2009-06-15</date><risdate>2009</risdate><volume>180</volume><issue>2</issue><spage>224</spage><epage>233</epage><pages>224-233</pages><issn>0165-0270</issn><eissn>1872-678X</eissn><abstract>Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100 fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion.</abstract><cop>Netherlands</cop><pub>Elsevier B.V</pub><pmid>19464514</pmid><doi>10.1016/j.jneumeth.2009.03.010</doi><tpages>10</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0165-0270
ispartof Journal of neuroscience methods, 2009-06, Vol.180 (2), p.224-233
issn 0165-0270
1872-678X
language eng
recordid cdi_proquest_miscellaneous_67275955
source MEDLINE; Elsevier ScienceDirect Journals
subjects Action Potentials - physiology
Algorithms
Animal movement
Animal posture
Animals
Biomechanical Phenomena - physiology
Brain - physiology
Brain–machine interface
Computer Simulation
Extremities - physiology
Image Processing, Computer-Assisted - instrumentation
Image Processing, Computer-Assisted - methods
Macaca mulatta
Motor Cortex - physiology
Movement - physiology
Neurons - physiology
Neurophysiology - instrumentation
Neurophysiology - methods
Pattern Recognition, Automated - methods
Primates
Range of Motion, Articular - physiology
Real-time systems
Signal Processing, Computer-Assisted
Time Factors
User-Computer Interface
Video Recording - instrumentation
Video Recording - methods
Video tracking
title Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-11T07%3A57%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Three-dimensional,%20automated,%20real-time%20video%20system%20for%20tracking%20limb%20motion%20in%20brain%E2%80%93machine%20interface%20studies&rft.jtitle=Journal%20of%20neuroscience%20methods&rft.au=Peikon,%20Ian%20D.&rft.date=2009-06-15&rft.volume=180&rft.issue=2&rft.spage=224&rft.epage=233&rft.pages=224-233&rft.issn=0165-0270&rft.eissn=1872-678X&rft_id=info:doi/10.1016/j.jneumeth.2009.03.010&rft_dat=%3Cproquest_cross%3E20079831%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=20079831&rft_id=info:pmid/19464514&rft_els_id=S0165027009001605&rfr_iscdi=true