University of Washington Indoor Object Manipulation (UW IOM) Dataset

The University of Washington Indoor Object Manipulation (UW IOM) dataset comprises videos (and corresponding skeletal tracking information) of twenty participants within the age group of 18-25 years, of which fifteen are males and the remaining five are females. The videos are recorded using a Kinec...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
1. Verfasser: Parsa, Behnoosh
Format: Dataset
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Parsa, Behnoosh
description The University of Washington Indoor Object Manipulation (UW IOM) dataset comprises videos (and corresponding skeletal tracking information) of twenty participants within the age group of 18-25 years, of which fifteen are males and the remaining five are females. The videos are recorded using a Kinect Sensor for Xbox One at an average rate of twelve frames per second. Each participant carries out the same set of tasks in terms of picking up six objects (three identical empty boxes and three identical rods) from three different vertical racks, placing them on a table, putting them back on the racks from where they are picked up, and then walking out of the scene carrying the box from the middle rack. The boxes are manipulated with both the hands while the rods are manipulated using only one hand. The above tasks are repeated in the same sequence three times such that the duration of every video is approximately three minutes. We categorize the actions into seventeen labels, where each label follows a four-tier hierarchy. The first tier indicates whether the box or the rod is manipulated, the second tier denotes human motion (walk, stand, and bend), the third tier captures the type of object manipulation if applicable (reach, pick-up, place, and hold), and the fourth tier represents the relative height of the surface where manipulation is taking place (low, medium, and high).
doi_str_mv 10.17632/xwzzkxtf9s
format Dataset
fullrecord <record><control><sourceid>datacite_PQ8</sourceid><recordid>TN_cdi_datacite_primary_10_17632_xwzzkxtf9s</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_17632_xwzzkxtf9s</sourcerecordid><originalsourceid>FETCH-datacite_primary_10_17632_xwzzkxtf9s3</originalsourceid><addsrcrecordid>eNpjYBA2NNAzNDczNtKvKK-qyq4oSbMs5mRwCc3LLEstKs4sqVTIT1MITyzOyMxLL8nPU_DMS8nPL1LwT8pKTS5R8E3MyywozUksyQRKaYSGK3j6-2oquCSWJBanlvAwsKYl5hSn8kJpbgZtN9cQZw_dFKB8cmZJanxBUWZuYlFlvKFBPNgN8Qg3GJOmGgCQcEIZ</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>dataset</recordtype></control><display><type>dataset</type><title>University of Washington Indoor Object Manipulation (UW IOM) Dataset</title><source>DataCite</source><creator>Parsa, Behnoosh</creator><creatorcontrib>Parsa, Behnoosh</creatorcontrib><description>The University of Washington Indoor Object Manipulation (UW IOM) dataset comprises videos (and corresponding skeletal tracking information) of twenty participants within the age group of 18-25 years, of which fifteen are males and the remaining five are females. The videos are recorded using a Kinect Sensor for Xbox One at an average rate of twelve frames per second. Each participant carries out the same set of tasks in terms of picking up six objects (three identical empty boxes and three identical rods) from three different vertical racks, placing them on a table, putting them back on the racks from where they are picked up, and then walking out of the scene carrying the box from the middle rack. The boxes are manipulated with both the hands while the rods are manipulated using only one hand. The above tasks are repeated in the same sequence three times such that the duration of every video is approximately three minutes. We categorize the actions into seventeen labels, where each label follows a four-tier hierarchy. The first tier indicates whether the box or the rod is manipulated, the second tier denotes human motion (walk, stand, and bend), the third tier captures the type of object manipulation if applicable (reach, pick-up, place, and hold), and the fourth tier represents the relative height of the surface where manipulation is taking place (low, medium, and high).</description><identifier>DOI: 10.17632/xwzzkxtf9s</identifier><language>eng</language><publisher>Mendeley</publisher><subject>Activity Recognition ; Computer Vision ; Deep Learning ; Ergonomics ; Motion Capture ; Video Processing</subject><creationdate>2020</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,1892</link.rule.ids><linktorsrc>$$Uhttps://commons.datacite.org/doi.org/10.17632/xwzzkxtf9s$$EView_record_in_DataCite.org$$FView_record_in_$$GDataCite.org$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>Parsa, Behnoosh</creatorcontrib><title>University of Washington Indoor Object Manipulation (UW IOM) Dataset</title><description>The University of Washington Indoor Object Manipulation (UW IOM) dataset comprises videos (and corresponding skeletal tracking information) of twenty participants within the age group of 18-25 years, of which fifteen are males and the remaining five are females. The videos are recorded using a Kinect Sensor for Xbox One at an average rate of twelve frames per second. Each participant carries out the same set of tasks in terms of picking up six objects (three identical empty boxes and three identical rods) from three different vertical racks, placing them on a table, putting them back on the racks from where they are picked up, and then walking out of the scene carrying the box from the middle rack. The boxes are manipulated with both the hands while the rods are manipulated using only one hand. The above tasks are repeated in the same sequence three times such that the duration of every video is approximately three minutes. We categorize the actions into seventeen labels, where each label follows a four-tier hierarchy. The first tier indicates whether the box or the rod is manipulated, the second tier denotes human motion (walk, stand, and bend), the third tier captures the type of object manipulation if applicable (reach, pick-up, place, and hold), and the fourth tier represents the relative height of the surface where manipulation is taking place (low, medium, and high).</description><subject>Activity Recognition</subject><subject>Computer Vision</subject><subject>Deep Learning</subject><subject>Ergonomics</subject><subject>Motion Capture</subject><subject>Video Processing</subject><fulltext>true</fulltext><rsrctype>dataset</rsrctype><creationdate>2020</creationdate><recordtype>dataset</recordtype><sourceid>PQ8</sourceid><recordid>eNpjYBA2NNAzNDczNtKvKK-qyq4oSbMs5mRwCc3LLEstKs4sqVTIT1MITyzOyMxLL8nPU_DMS8nPL1LwT8pKTS5R8E3MyywozUksyQRKaYSGK3j6-2oquCSWJBanlvAwsKYl5hSn8kJpbgZtN9cQZw_dFKB8cmZJanxBUWZuYlFlvKFBPNgN8Qg3GJOmGgCQcEIZ</recordid><startdate>20201102</startdate><enddate>20201102</enddate><creator>Parsa, Behnoosh</creator><general>Mendeley</general><scope>DYCCY</scope><scope>PQ8</scope></search><sort><creationdate>20201102</creationdate><title>University of Washington Indoor Object Manipulation (UW IOM) Dataset</title><author>Parsa, Behnoosh</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-datacite_primary_10_17632_xwzzkxtf9s3</frbrgroupid><rsrctype>datasets</rsrctype><prefilter>datasets</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Activity Recognition</topic><topic>Computer Vision</topic><topic>Deep Learning</topic><topic>Ergonomics</topic><topic>Motion Capture</topic><topic>Video Processing</topic><toplevel>online_resources</toplevel><creatorcontrib>Parsa, Behnoosh</creatorcontrib><collection>DataCite (Open Access)</collection><collection>DataCite</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Parsa, Behnoosh</au><format>book</format><genre>unknown</genre><ristype>DATA</ristype><title>University of Washington Indoor Object Manipulation (UW IOM) Dataset</title><date>2020-11-02</date><risdate>2020</risdate><abstract>The University of Washington Indoor Object Manipulation (UW IOM) dataset comprises videos (and corresponding skeletal tracking information) of twenty participants within the age group of 18-25 years, of which fifteen are males and the remaining five are females. The videos are recorded using a Kinect Sensor for Xbox One at an average rate of twelve frames per second. Each participant carries out the same set of tasks in terms of picking up six objects (three identical empty boxes and three identical rods) from three different vertical racks, placing them on a table, putting them back on the racks from where they are picked up, and then walking out of the scene carrying the box from the middle rack. The boxes are manipulated with both the hands while the rods are manipulated using only one hand. The above tasks are repeated in the same sequence three times such that the duration of every video is approximately three minutes. We categorize the actions into seventeen labels, where each label follows a four-tier hierarchy. The first tier indicates whether the box or the rod is manipulated, the second tier denotes human motion (walk, stand, and bend), the third tier captures the type of object manipulation if applicable (reach, pick-up, place, and hold), and the fourth tier represents the relative height of the surface where manipulation is taking place (low, medium, and high).</abstract><pub>Mendeley</pub><doi>10.17632/xwzzkxtf9s</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.17632/xwzzkxtf9s
ispartof
issn
language eng
recordid cdi_datacite_primary_10_17632_xwzzkxtf9s
source DataCite
subjects Activity Recognition
Computer Vision
Deep Learning
Ergonomics
Motion Capture
Video Processing
title University of Washington Indoor Object Manipulation (UW IOM) Dataset
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-10T06%3A20%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-datacite_PQ8&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=unknown&rft.au=Parsa,%20Behnoosh&rft.date=2020-11-02&rft_id=info:doi/10.17632/xwzzkxtf9s&rft_dat=%3Cdatacite_PQ8%3E10_17632_xwzzkxtf9s%3C/datacite_PQ8%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true