Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies
Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb...
Gespeichert in:
Veröffentlicht in: | Journal of neuroscience methods 2009-06, Vol.180 (2), p.224-233 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 233 |
---|---|
container_issue | 2 |
container_start_page | 224 |
container_title | Journal of neuroscience methods |
container_volume | 180 |
creator | Peikon, Ian D. Fitzsimmons, Nathan A. Lebedev, Mikhail A. Nicolelis, Miguel A.L. |
description | Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100
fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion. |
doi_str_mv | 10.1016/j.jneumeth.2009.03.010 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_67275955</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0165027009001605</els_id><sourcerecordid>20079831</sourcerecordid><originalsourceid>FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</originalsourceid><addsrcrecordid>eNqFkcFu1DAQhi1ERZfCK1Q-cWrCOHac9Q1U0YJUiUuRuFmOPWG9xE6xnUq98Q68IU9SV7uIY08j_fP9M5r5CTln0DJg8v2-3UdcA5Zd2wGoFngLDF6QDdsOXSOH7feXZFPBvoFugFPyOuc9AAgF8hU5ZUpI0TOxIel2lxAb5wPG7Jdo5gtq1rIEU9Bd0IRmbkpt0nvvcKH5IRcMdFoSLcnYnz7-oLMPIw1LqW7qIx2T8fHv7z_B2J2PWKWCaTIWaS6r85jfkJPJzBnfHusZ-Xb16fbyc3Pz9frL5cebxnI1lMYZLkANCKMAxiWK0Y3jhF03TY4LLl0_WgnKwWAFYD2s722nqu46y0cm-Bl5d5h7l5ZfK-aig88W59lEXNas5dANver7Z8H64EFtOaugPIA2LTknnPRd8sGkB81AP8Wi9_pfLE8upYHrGks1nh83rGNA9992zKECHw4A1ofce0w6W4_RovMJbdFu8c_teASojqTs</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>20079831</pqid></control><display><type>article</type><title>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</title><source>MEDLINE</source><source>Elsevier ScienceDirect Journals</source><creator>Peikon, Ian D. ; Fitzsimmons, Nathan A. ; Lebedev, Mikhail A. ; Nicolelis, Miguel A.L.</creator><creatorcontrib>Peikon, Ian D. ; Fitzsimmons, Nathan A. ; Lebedev, Mikhail A. ; Nicolelis, Miguel A.L.</creatorcontrib><description>Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100
fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion.</description><identifier>ISSN: 0165-0270</identifier><identifier>EISSN: 1872-678X</identifier><identifier>DOI: 10.1016/j.jneumeth.2009.03.010</identifier><identifier>PMID: 19464514</identifier><language>eng</language><publisher>Netherlands: Elsevier B.V</publisher><subject>Action Potentials - physiology ; Algorithms ; Animal movement ; Animal posture ; Animals ; Biomechanical Phenomena - physiology ; Brain - physiology ; Brain–machine interface ; Computer Simulation ; Extremities - physiology ; Image Processing, Computer-Assisted - instrumentation ; Image Processing, Computer-Assisted - methods ; Macaca mulatta ; Motor Cortex - physiology ; Movement - physiology ; Neurons - physiology ; Neurophysiology - instrumentation ; Neurophysiology - methods ; Pattern Recognition, Automated - methods ; Primates ; Range of Motion, Articular - physiology ; Real-time systems ; Signal Processing, Computer-Assisted ; Time Factors ; User-Computer Interface ; Video Recording - instrumentation ; Video Recording - methods ; Video tracking</subject><ispartof>Journal of neuroscience methods, 2009-06, Vol.180 (2), p.224-233</ispartof><rights>2009 Elsevier B.V.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</citedby><cites>FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S0165027009001605$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65306</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/19464514$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Peikon, Ian D.</creatorcontrib><creatorcontrib>Fitzsimmons, Nathan A.</creatorcontrib><creatorcontrib>Lebedev, Mikhail A.</creatorcontrib><creatorcontrib>Nicolelis, Miguel A.L.</creatorcontrib><title>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</title><title>Journal of neuroscience methods</title><addtitle>J Neurosci Methods</addtitle><description>Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100
fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion.</description><subject>Action Potentials - physiology</subject><subject>Algorithms</subject><subject>Animal movement</subject><subject>Animal posture</subject><subject>Animals</subject><subject>Biomechanical Phenomena - physiology</subject><subject>Brain - physiology</subject><subject>Brain–machine interface</subject><subject>Computer Simulation</subject><subject>Extremities - physiology</subject><subject>Image Processing, Computer-Assisted - instrumentation</subject><subject>Image Processing, Computer-Assisted - methods</subject><subject>Macaca mulatta</subject><subject>Motor Cortex - physiology</subject><subject>Movement - physiology</subject><subject>Neurons - physiology</subject><subject>Neurophysiology - instrumentation</subject><subject>Neurophysiology - methods</subject><subject>Pattern Recognition, Automated - methods</subject><subject>Primates</subject><subject>Range of Motion, Articular - physiology</subject><subject>Real-time systems</subject><subject>Signal Processing, Computer-Assisted</subject><subject>Time Factors</subject><subject>User-Computer Interface</subject><subject>Video Recording - instrumentation</subject><subject>Video Recording - methods</subject><subject>Video tracking</subject><issn>0165-0270</issn><issn>1872-678X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2009</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNqFkcFu1DAQhi1ERZfCK1Q-cWrCOHac9Q1U0YJUiUuRuFmOPWG9xE6xnUq98Q68IU9SV7uIY08j_fP9M5r5CTln0DJg8v2-3UdcA5Zd2wGoFngLDF6QDdsOXSOH7feXZFPBvoFugFPyOuc9AAgF8hU5ZUpI0TOxIel2lxAb5wPG7Jdo5gtq1rIEU9Bd0IRmbkpt0nvvcKH5IRcMdFoSLcnYnz7-oLMPIw1LqW7qIx2T8fHv7z_B2J2PWKWCaTIWaS6r85jfkJPJzBnfHusZ-Xb16fbyc3Pz9frL5cebxnI1lMYZLkANCKMAxiWK0Y3jhF03TY4LLl0_WgnKwWAFYD2s722nqu46y0cm-Bl5d5h7l5ZfK-aig88W59lEXNas5dANver7Z8H64EFtOaugPIA2LTknnPRd8sGkB81AP8Wi9_pfLE8upYHrGks1nh83rGNA9992zKECHw4A1ofce0w6W4_RovMJbdFu8c_teASojqTs</recordid><startdate>20090615</startdate><enddate>20090615</enddate><creator>Peikon, Ian D.</creator><creator>Fitzsimmons, Nathan A.</creator><creator>Lebedev, Mikhail A.</creator><creator>Nicolelis, Miguel A.L.</creator><general>Elsevier B.V</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7TK</scope><scope>7X8</scope></search><sort><creationdate>20090615</creationdate><title>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</title><author>Peikon, Ian D. ; Fitzsimmons, Nathan A. ; Lebedev, Mikhail A. ; Nicolelis, Miguel A.L.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c397t-da34097e0b40136e4bdbbfe22ffd3436d5bc609d07c40e49055c2936dd2c3b143</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2009</creationdate><topic>Action Potentials - physiology</topic><topic>Algorithms</topic><topic>Animal movement</topic><topic>Animal posture</topic><topic>Animals</topic><topic>Biomechanical Phenomena - physiology</topic><topic>Brain - physiology</topic><topic>Brain–machine interface</topic><topic>Computer Simulation</topic><topic>Extremities - physiology</topic><topic>Image Processing, Computer-Assisted - instrumentation</topic><topic>Image Processing, Computer-Assisted - methods</topic><topic>Macaca mulatta</topic><topic>Motor Cortex - physiology</topic><topic>Movement - physiology</topic><topic>Neurons - physiology</topic><topic>Neurophysiology - instrumentation</topic><topic>Neurophysiology - methods</topic><topic>Pattern Recognition, Automated - methods</topic><topic>Primates</topic><topic>Range of Motion, Articular - physiology</topic><topic>Real-time systems</topic><topic>Signal Processing, Computer-Assisted</topic><topic>Time Factors</topic><topic>User-Computer Interface</topic><topic>Video Recording - instrumentation</topic><topic>Video Recording - methods</topic><topic>Video tracking</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Peikon, Ian D.</creatorcontrib><creatorcontrib>Fitzsimmons, Nathan A.</creatorcontrib><creatorcontrib>Lebedev, Mikhail A.</creatorcontrib><creatorcontrib>Nicolelis, Miguel A.L.</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Neurosciences Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>Journal of neuroscience methods</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Peikon, Ian D.</au><au>Fitzsimmons, Nathan A.</au><au>Lebedev, Mikhail A.</au><au>Nicolelis, Miguel A.L.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies</atitle><jtitle>Journal of neuroscience methods</jtitle><addtitle>J Neurosci Methods</addtitle><date>2009-06-15</date><risdate>2009</risdate><volume>180</volume><issue>2</issue><spage>224</spage><epage>233</epage><pages>224-233</pages><issn>0165-0270</issn><eissn>1872-678X</eissn><abstract>Collection and analysis of limb kinematic data are essential components of the study of biological motion, including research into biomechanics, kinesiology, neurophysiology and brain–machine interfaces (BMIs). In particular, BMI research requires advanced, real-time systems capable of sampling limb kinematics with minimal contact to the subject's body. To answer this demand, we have developed an automated video tracking system for real-time tracking of multiple body parts in freely behaving primates. The system employs high-contrast markers painted on the animal's joints to continuously track the three-dimensional positions of their limbs during activity. Two-dimensional coordinates captured by each video camera are combined and converted to three-dimensional coordinates using a quadratic fitting algorithm. Real-time operation of the system is accomplished using direct memory access (DMA). The system tracks the markers at a rate of 52 frames per second (fps) in real-time and up to 100
fps if video recordings are captured to be later analyzed off-line. The system has been tested in several BMI primate experiments, in which limb position was sampled simultaneously with chronic recordings of the extracellular activity of hundreds of cortical cells. During these recordings, multiple computational models were employed to extract a series of kinematic parameters from neuronal ensemble activity in real-time. The system operated reliably under these experimental conditions and was able to compensate for marker occlusions that occurred during natural movements. We propose that this system could also be extended to applications that include other classes of biological motion.</abstract><cop>Netherlands</cop><pub>Elsevier B.V</pub><pmid>19464514</pmid><doi>10.1016/j.jneumeth.2009.03.010</doi><tpages>10</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0165-0270 |
ispartof | Journal of neuroscience methods, 2009-06, Vol.180 (2), p.224-233 |
issn | 0165-0270 1872-678X |
language | eng |
recordid | cdi_proquest_miscellaneous_67275955 |
source | MEDLINE; Elsevier ScienceDirect Journals |
subjects | Action Potentials - physiology Algorithms Animal movement Animal posture Animals Biomechanical Phenomena - physiology Brain - physiology Brain–machine interface Computer Simulation Extremities - physiology Image Processing, Computer-Assisted - instrumentation Image Processing, Computer-Assisted - methods Macaca mulatta Motor Cortex - physiology Movement - physiology Neurons - physiology Neurophysiology - instrumentation Neurophysiology - methods Pattern Recognition, Automated - methods Primates Range of Motion, Articular - physiology Real-time systems Signal Processing, Computer-Assisted Time Factors User-Computer Interface Video Recording - instrumentation Video Recording - methods Video tracking |
title | Three-dimensional, automated, real-time video system for tracking limb motion in brain–machine interface studies |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-11T07%3A57%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Three-dimensional,%20automated,%20real-time%20video%20system%20for%20tracking%20limb%20motion%20in%20brain%E2%80%93machine%20interface%20studies&rft.jtitle=Journal%20of%20neuroscience%20methods&rft.au=Peikon,%20Ian%20D.&rft.date=2009-06-15&rft.volume=180&rft.issue=2&rft.spage=224&rft.epage=233&rft.pages=224-233&rft.issn=0165-0270&rft.eissn=1872-678X&rft_id=info:doi/10.1016/j.jneumeth.2009.03.010&rft_dat=%3Cproquest_cross%3E20079831%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=20079831&rft_id=info:pmid/19464514&rft_els_id=S0165027009001605&rfr_iscdi=true |