Annotation of 3D Object Pose from Egocentric Video
Authors
Published Date
Publisher
Abstract
Current datasets for human egocentric vision have been focused on 2D tasks such as object segmentation, object detection, and activity recognition. For extended tasks such as AR and robotics, 3D understanding is needed to learn human skills. A new dataset called Indoor 3D Egocentric Object (IDEO) was created to address the shortcomings of existing datasets. Part of this dataset includes annotated 9-DOF pose of objects being manipulated by hand, which introduces nontrivial characteristic occlusion and uncommon object poses. A novel annotation tool was created to streamline the annotation process allowing for fast and accurate annotation by untrained crowd workers for the 9-DOF pose (translation, rotation, scale) of objects being manipulated by hand in egocentric images.
Keywords
Description
University of Minnesota M.S. thesis. April 2023. Major: Computer Science. Advisor: Hyun Soo Park. 1 computer file (PDF); v, 34 pages.
Related to
item.page.replaces
License
Series/Report Number
Funding Information
item.page.isbn
DOI identifier
Previously Published Citation
Other identifiers
Suggested Citation
Lemke, Lance. (2023). Annotation of 3D Object Pose from Egocentric Video. Retrieved from the University Digital Conservancy, https://hdl.handle.net/11299/256957.
Content distributed via the University Digital Conservancy may be subject to additional license and use restrictions applied by the depositor. By using these files, users agree to the Terms of Use. Materials in the UDC may contain content that is disturbing and/or harmful. For more information, please see our statement on harmful content in digital repositories.
