data
dict
{ "proceeding": { "id": "1CJbEwHHqEg", "title": "2022 IEEE on Conference Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1CJc5J6RYYM", "doi": "10.1109/VR51125.2022.00035", "title": "Foldable Spaces: An Overt Redirection Approach for Natural Walking in Virtual Reality", "normalizedTitle": "Foldable Spaces: An Overt Redirection Approach for Natural Walking in Virtual Reality", "abstract": "Overt redirection is a class of virtual reality locomotion that uses perceptible transformations to enable the user to naturally walk through a virtual environment larger than the physical tracking space. In this research, we propose Foldable Spaces, a novel redirection approach based on the idea of dynamically ‘folding’ the geometry of the virtual environment to reveal new locations depending on the trajectory of the virtual reality user. Based on this approach, we developed three distinct techniques for overt redirection: (1) Horizontal, which folds and reveals layers of virtual space like the pages in a book; (2) Vertical, which rotates virtual space towards the user along a vertical axis; and (3) Accordion, which corrugates and flattens virtual space to bring faraway places closer to the user. In a within-subjects user study, we compared our proposed foldable techniques against each other along with a similarly situated redirection technique, Stop & Reset. Our findings show that Accordion was the most well-received by participants in terms of providing a smooth, continuous, and ‘natural’ experience of walking that does not involve shifts in orientation and provides an overarching view through the virtual environment.", "abstracts": [ { "abstractType": "Regular", "content": "Overt redirection is a class of virtual reality locomotion that uses perceptible transformations to enable the user to naturally walk through a virtual environment larger than the physical tracking space. In this research, we propose Foldable Spaces, a novel redirection approach based on the idea of dynamically ‘folding’ the geometry of the virtual environment to reveal new locations depending on the trajectory of the virtual reality user. Based on this approach, we developed three distinct techniques for overt redirection: (1) Horizontal, which folds and reveals layers of virtual space like the pages in a book; (2) Vertical, which rotates virtual space towards the user along a vertical axis; and (3) Accordion, which corrugates and flattens virtual space to bring faraway places closer to the user. In a within-subjects user study, we compared our proposed foldable techniques against each other along with a similarly situated redirection technique, Stop & Reset. Our findings show that Accordion was the most well-received by participants in terms of providing a smooth, continuous, and ‘natural’ experience of walking that does not involve shifts in orientation and provides an overarching view through the virtual environment.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Overt redirection is a class of virtual reality locomotion that uses perceptible transformations to enable the user to naturally walk through a virtual environment larger than the physical tracking space. In this research, we propose Foldable Spaces, a novel redirection approach based on the idea of dynamically ‘folding’ the geometry of the virtual environment to reveal new locations depending on the trajectory of the virtual reality user. Based on this approach, we developed three distinct techniques for overt redirection: (1) Horizontal, which folds and reveals layers of virtual space like the pages in a book; (2) Vertical, which rotates virtual space towards the user along a vertical axis; and (3) Accordion, which corrugates and flattens virtual space to bring faraway places closer to the user. In a within-subjects user study, we compared our proposed foldable techniques against each other along with a similarly situated redirection technique, Stop & Reset. Our findings show that Accordion was the most well-received by participants in terms of providing a smooth, continuous, and ‘natural’ experience of walking that does not involve shifts in orientation and provides an overarching view through the virtual environment.", "fno": "961700a167", "keywords": [ "User Interfaces", "Virtual Reality", "Foldable Spaces", "Redirection Technique", "Foldable Techniques", "Within Subjects User Study", "Virtual Space", "Virtual Reality User", "Redirection Approach", "Physical Tracking Space", "Virtual Environment", "Virtual Reality Locomotion", "Natural Walking", "Overt Redirection Approach", "Legged Locomotion", "Geometry", "Three Dimensional Displays", "Conferences", "Virtual Environments", "Medical Treatment", "User Interfaces", "Human Centered Computing", "Interaction Paradigms", "Virtual Reality" ], "authors": [ { "affiliation": "KU Leuven", "fullName": "Jihae Han", "givenName": "Jihae", "surname": "Han", "__typename": "ArticleAuthorType" }, { "affiliation": "KU Leuven", "fullName": "Andrew Vande Moere", "givenName": "Andrew Vande", "surname": "Moere", "__typename": "ArticleAuthorType" }, { "affiliation": "KU Leuven", "fullName": "Adalberto L. Simeone", "givenName": "Adalberto L.", "surname": "Simeone", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-03-01T00:00:00", "pubType": "proceedings", "pages": "167-175", "year": "2022", "issn": null, "isbn": "978-1-6654-9617-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1CJc5oVE9Co", "name": "pvr202296170-09756751s1-mm_961700a167.zip", "size": "203 MB", "location": "https://www.computer.org/csdl/api/v1/extra/pvr202296170-09756751s1-mm_961700a167.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "961700a158", "articleId": "1CJbPaYWLKM", "__typename": "AdjacentArticleType" }, "next": { "fno": "961700a176", "articleId": "1CJccMXHjKU", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2016/0836/0/07504752", "title": "Disguising rotational gain for redirected walking in virtual reality: Effect of visual density", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504752/12OmNyr8YkS", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2016/0836/0/07504714", "title": "A realistic walking model for enhancing redirection in virtual reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504714/12OmNzaQowA", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/05/09715723", "title": "Adaptive Redirection: A Context-Aware Redirected Walking Meta-Strategy", "doi": null, "abstractUrl": "/journal/tg/2022/05/09715723/1B4hxCQXB4c", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2022/9617/0/961700a464", "title": "RedirectedDoors: Redirection While Opening Doors in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2022/961700a464/1CJc9xfqBSo", "parentPublication": { "id": "proceedings/vr/2022/9617/0", "title": "2022 IEEE on Conference Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09785918", "title": "Redirected Walking for Exploring Immersive Virtual Spaces with HMD: A Comprehensive Review and Recent Advances", "doi": null, "abstractUrl": "/journal/tg/5555/01/09785918/1DPaEdHg6KQ", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09893374", "title": "A Segmented Redirection Mapping Method for Roadmaps of Large Constrained Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/5555/01/09893374/1GGLIh8KmSA", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798143", "title": "Estimating Detection Thresholds for Desktop-Scale Hand Redirection in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798143/1cJ0GRxSQwM", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089579", "title": "Feature Guided Path Redirection for VR Navigation", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089579/1jIx7XMm676", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089543", "title": "The Space Bender: Supporting Natural Walking via Overt Manipulation of the Virtual Environment", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089543/1jIxb9EbKJq", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/05/09382909", "title": "ARC: Alignment-based Redirection Controller for Redirected Walking in Complex Environments", "doi": null, "abstractUrl": "/journal/tg/2021/05/09382909/1saZt58Vwf6", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1CJcAaH6aYg", "title": "2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "acronym": "vrw", "groupId": "1836626", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1CJe2Dfxj8Y", "doi": "10.1109/VRW55335.2022.00162", "title": "Minimaps for Impossible Spaces: Improving Spatial Cognition in Self-Overlapping Virtual Rooms", "normalizedTitle": "Minimaps for Impossible Spaces: Improving Spatial Cognition in Self-Overlapping Virtual Rooms", "abstract": "Natural walking in virtual reality is constrained by the physical boundaries defined by the size of the user's tracking space. Impossible spaces, a redirected walking technique, enlarge the virtual environment by creating overlapping architecture and letting multiple locations occupy the same physical space. Minimaps, which are small representations of the user's surroundings, are a common method to assist with wayfinding and navigation. Unfortunately, in a naive implementation of such minimaps for an environment with impossible spaces, the overlap would be obvious. We investigated approaches to displaying impossible spaces on minimaps without attracting users' attention to the overlapping parts of the VE. In this paper, we present a study that compared spatial cognition in a short gamified VR experience between three different conditions: (i) without minimap, (ii) with basic minimap, and (iii) with full minimap. Our results show significant improvements on accuracy in direction estimation tasks, suggesting that minimaps help users acquire spatial knowledge in the presence of impossible spaces.", "abstracts": [ { "abstractType": "Regular", "content": "Natural walking in virtual reality is constrained by the physical boundaries defined by the size of the user's tracking space. Impossible spaces, a redirected walking technique, enlarge the virtual environment by creating overlapping architecture and letting multiple locations occupy the same physical space. Minimaps, which are small representations of the user's surroundings, are a common method to assist with wayfinding and navigation. Unfortunately, in a naive implementation of such minimaps for an environment with impossible spaces, the overlap would be obvious. We investigated approaches to displaying impossible spaces on minimaps without attracting users' attention to the overlapping parts of the VE. In this paper, we present a study that compared spatial cognition in a short gamified VR experience between three different conditions: (i) without minimap, (ii) with basic minimap, and (iii) with full minimap. Our results show significant improvements on accuracy in direction estimation tasks, suggesting that minimaps help users acquire spatial knowledge in the presence of impossible spaces.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Natural walking in virtual reality is constrained by the physical boundaries defined by the size of the user's tracking space. Impossible spaces, a redirected walking technique, enlarge the virtual environment by creating overlapping architecture and letting multiple locations occupy the same physical space. Minimaps, which are small representations of the user's surroundings, are a common method to assist with wayfinding and navigation. Unfortunately, in a naive implementation of such minimaps for an environment with impossible spaces, the overlap would be obvious. We investigated approaches to displaying impossible spaces on minimaps without attracting users' attention to the overlapping parts of the VE. In this paper, we present a study that compared spatial cognition in a short gamified VR experience between three different conditions: (i) without minimap, (ii) with basic minimap, and (iii) with full minimap. Our results show significant improvements on accuracy in direction estimation tasks, suggesting that minimaps help users acquire spatial knowledge in the presence of impossible spaces.", "fno": "840200a622", "keywords": [ "Cognition", "Virtual Reality", "Overlapping Architecture", "Physical Space", "Minimap", "Spatial Cognition", "Self Overlapping Virtual Rooms", "Virtual Reality", "Virtual Environment", "Legged Locomotion", "Three Dimensional Displays", "Navigation", "Conferences", "Virtual Environments", "Estimation", "Computer Architecture", "Human Centered Computing X 2014 Human Computer Interaction HCI X 2014 Interaction Paradigms X 2014 Virtual Reality" ], "authors": [ { "affiliation": "Universität Hamburg", "fullName": "Rafael Epplée", "givenName": "Rafael", "surname": "Epplée", "__typename": "ArticleAuthorType" }, { "affiliation": "HAW Hamburg", "fullName": "Eike Langbehn", "givenName": "Eike", "surname": "Langbehn", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-03-01T00:00:00", "pubType": "proceedings", "pages": "622-623", "year": "2022", "issn": null, "isbn": "978-1-6654-8402-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "840200a620", "articleId": "1CJdavNhwAw", "__typename": "AdjacentArticleType" }, "next": { "fno": "840200a624", "articleId": "1CJdePZnsfS", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2017/6647/0/07892335", "title": "Designing intentional impossible spaces in virtual reality narratives: A case study", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892335/12OmNApcu9b", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2013/4795/0/06549386", "title": "Flexible spaces: A virtual step outside of reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549386/12OmNBOllfZ", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2017/6647/0/07892303", "title": "Object location memory error in virtual and real environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892303/12OmNx7ouWn", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2013/6097/0/06550194", "title": "Flexible spaces: Dynamic layout generation for infinite walking in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2013/06550194/12OmNyFU75b", "parentPublication": { "id": "proceedings/3dui/2013/6097/0", "title": "2013 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446579", "title": "Leveraging Configuration Spaces and Navigation Functions for Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446579/13bd1fdV4lq", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446167", "title": "Redirected Spaces: Going Beyond Borders", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446167/13bd1fph1xv", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040555", "title": "Impossible Spaces: Maximizing Natural Walking in Virtual Environments with Self-Overlapping Architecture", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040555/13rRUygBwhF", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798319", "title": "Simulation and Evaluation of Three-User Redirected Walking Algorithm in Shared Physical Spaces", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798319/1cJ1aPwr8l2", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090660", "title": "Relative Room Size Judgments in Impossible Spaces", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090660/1jIxqsUN6ik", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2021/4057/0/405700a389", "title": "Impossible Open Spaces: Exploring the Effects of Occlusion on the Noticeability of Self-Overlapping Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/vrw/2021/405700a389/1tnWx88Rxuw", "parentPublication": { "id": "proceedings/vrw/2021/4057/0", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1cI6akLvAuQ", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1cJ0OPBhW4U", "doi": "10.1109/VR.2019.8798074", "title": "VRoamer: Generating On-The-Fly VR Experiences While Walking inside Large, Unknown Real-World Building Environments", "normalizedTitle": "VRoamer: Generating On-The-Fly VR Experiences While Walking inside Large, Unknown Real-World Building Environments", "abstract": "Procedural generation in virtual reality (VR) has been used to adapt the virtual world to various indoor environments, fitting different geometries and interiors with virtual environments. However, such applications require that the physical environment be known or pre-scanned prior to use to then generate the corresponding virtual scene, thus restricting the virtual experience to a controlled space. In this paper, we present VRoamer, which enables users to walk unseen physical spaces for which VRoamer procedurally generates a virtual scene on-the-fly. Scaling to the size of office buildings, VRoamer extracts walkable areas and detects physical obstacles in real time, instantiates pre-authored virtual rooms if their sizes fit physically walkable areas or otherwise generates virtual corridors and doors that lead to undiscovered physical areas. The use of these virtual structures allows VRoamer to (1) temporarily block users' passage, thus slowing them down while increasing VRoamer's insight into newly discovered physical areas, (2) prevent users from seeing changes beyond the current virtual scene, and (3) obfuscate the appearance of physical environments. VRoamer animates virtual objects to reflect dynamically discovered changes of the physical environment, such as people walking by or obstacles that become apparent. In our proof-of-concept study, participants were able to walk long distances through a procedurally generated dungeon experience and reported high levels of immersion.", "abstracts": [ { "abstractType": "Regular", "content": "Procedural generation in virtual reality (VR) has been used to adapt the virtual world to various indoor environments, fitting different geometries and interiors with virtual environments. However, such applications require that the physical environment be known or pre-scanned prior to use to then generate the corresponding virtual scene, thus restricting the virtual experience to a controlled space. In this paper, we present VRoamer, which enables users to walk unseen physical spaces for which VRoamer procedurally generates a virtual scene on-the-fly. Scaling to the size of office buildings, VRoamer extracts walkable areas and detects physical obstacles in real time, instantiates pre-authored virtual rooms if their sizes fit physically walkable areas or otherwise generates virtual corridors and doors that lead to undiscovered physical areas. The use of these virtual structures allows VRoamer to (1) temporarily block users' passage, thus slowing them down while increasing VRoamer's insight into newly discovered physical areas, (2) prevent users from seeing changes beyond the current virtual scene, and (3) obfuscate the appearance of physical environments. VRoamer animates virtual objects to reflect dynamically discovered changes of the physical environment, such as people walking by or obstacles that become apparent. In our proof-of-concept study, participants were able to walk long distances through a procedurally generated dungeon experience and reported high levels of immersion.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Procedural generation in virtual reality (VR) has been used to adapt the virtual world to various indoor environments, fitting different geometries and interiors with virtual environments. However, such applications require that the physical environment be known or pre-scanned prior to use to then generate the corresponding virtual scene, thus restricting the virtual experience to a controlled space. In this paper, we present VRoamer, which enables users to walk unseen physical spaces for which VRoamer procedurally generates a virtual scene on-the-fly. Scaling to the size of office buildings, VRoamer extracts walkable areas and detects physical obstacles in real time, instantiates pre-authored virtual rooms if their sizes fit physically walkable areas or otherwise generates virtual corridors and doors that lead to undiscovered physical areas. The use of these virtual structures allows VRoamer to (1) temporarily block users' passage, thus slowing them down while increasing VRoamer's insight into newly discovered physical areas, (2) prevent users from seeing changes beyond the current virtual scene, and (3) obfuscate the appearance of physical environments. VRoamer animates virtual objects to reflect dynamically discovered changes of the physical environment, such as people walking by or obstacles that become apparent. In our proof-of-concept study, participants were able to walk long distances through a procedurally generated dungeon experience and reported high levels of immersion.", "fno": "08798074", "keywords": [ "Computer Animation", "Computer Games", "Virtual Reality", "Indoor Environments", "Virtual Scene On The Fly", "Physically Walkable Areas", "Virtual Corridors", "V Roamer Animates Virtual Objects", "Physical Obstacles", "Office Buildings", "Legged Locomotion", "Geometry", "Virtual Environments", "Aerospace Electronics", "Real Time Systems", "Cameras", "Virtual Reality", "Procedural Generation", "Real Walking", "Locomotion Techniques", "Redirected Walking", "H 5 1 Information Interfaces And Presentation Multimedia Information Systems Virtual Realities" ], "authors": [ { "affiliation": "Microsoft Research Redmond", "fullName": "Lung-Pan Cheng", "givenName": "Lung-Pan", "surname": "Cheng", "__typename": "ArticleAuthorType" }, { "affiliation": "Microsoft Research Redmond", "fullName": "Eyal Ofek", "givenName": "Eyal", "surname": "Ofek", "__typename": "ArticleAuthorType" }, { "affiliation": "Microsoft Research Redmond", "fullName": "Christian Holz", "givenName": "Christian", "surname": "Holz", "__typename": "ArticleAuthorType" }, { "affiliation": "Microsoft Research Redmond", "fullName": "Andrew D. Wilson", "givenName": "Andrew D.", "surname": "Wilson", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-03-01T00:00:00", "pubType": "proceedings", "pages": "359-366", "year": "2019", "issn": null, "isbn": "978-1-7281-1377-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08798240", "articleId": "1cJ0UBOdi6I", "__typename": "AdjacentArticleType" }, "next": { "fno": "08797751", "articleId": "1cJ0WSuJ27e", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2016/0836/0/07504742", "title": "Simultaneous mapping and redirected walking for ad hoc free walking in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504742/12OmNyUFg0I", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pive/2012/1218/0/06229795", "title": "Adaptive redirected walking in a virtual world", "doi": null, "abstractUrl": "/proceedings-article/pive/2012/06229795/12OmNzUxOk4", "parentPublication": { "id": "proceedings/pive/2012/1218/0", "title": "2012 IEEE VR Workshop on Perceptual Illusions in Virtual Environments", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446216", "title": "I Can See on My Feet While Walking: Sensitivity to Translation Gains with Visible Feet", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446216/13bd1gJ1v0k", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/12/08067498", "title": "Oasis: Procedurally Generated Social Virtual Spaces from 3D Scanned Real Spaces", "doi": null, "abstractUrl": "/journal/tg/2018/12/08067498/14H4WRabB9m", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09733261", "title": "One-step out-of-place resetting for redirected walking in VR", "doi": null, "abstractUrl": "/journal/tg/5555/01/09733261/1BENJyPkx5S", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09785918", "title": "Redirected Walking for Exploring Immersive Virtual Spaces with HMD: A Comprehensive Review and Recent Advances", "doi": null, "abstractUrl": "/journal/tg/5555/01/09785918/1DPaEdHg6KQ", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2022/5325/0/532500a758", "title": "Infinite Virtual Space Exploration Using Space Tiling and Perceivable Reset at Fixed Positions", "doi": null, "abstractUrl": "/proceedings-article/ismar/2022/532500a758/1JrRneazFCw", "parentPublication": { "id": "proceedings/ismar/2022/5325/0", "title": "2022 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2023/4815/0/481500a053", "title": "Redirected Walking Based on Historical User Walking Data", "doi": null, "abstractUrl": "/proceedings-article/vr/2023/481500a053/1MNgUnNG7Ju", "parentPublication": { "id": "proceedings/vr/2023/4815/0", "title": "2023 IEEE Conference Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08967011", "title": "The Role of Binocular Vision in Avoiding Virtual Obstacles While Walking", "doi": null, "abstractUrl": "/journal/tg/2021/07/08967011/1gPjyDVBxF6", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/11/09523832", "title": "Redirected Walking in Static and Dynamic Scenes Using Visibility Polygons", "doi": null, "abstractUrl": "/journal/tg/2021/11/09523832/1wpqjiNuSqY", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1jIxhEnA8IE", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "acronym": "vrw", "groupId": "1836626", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1jIxqsUN6ik", "doi": "10.1109/VRW50115.2020.00171", "title": "Relative Room Size Judgments in Impossible Spaces", "normalizedTitle": "Relative Room Size Judgments in Impossible Spaces", "abstract": "In VR, “impossible spaces” allow virtual rooms to occupy overlapping portions of physical space. We investigated how size judgments of impossible spaces are influenced by the ratio of the sizes of overlapping rooms. Participants (n=36) were randomly assigned to one of three conditions and completed 13 trials. Participants’ reports accurately reflected the true ratio of the rooms in all conditions; however, participants reported less extreme ratios as the ratio increased. The results suggest that important spatial relationships are preserved in impossible spaces, namely 1) judged sizes of individual rooms, and 2) judgments concerning the relative sizes of different rooms.", "abstracts": [ { "abstractType": "Regular", "content": "In VR, “impossible spaces” allow virtual rooms to occupy overlapping portions of physical space. We investigated how size judgments of impossible spaces are influenced by the ratio of the sizes of overlapping rooms. Participants (n=36) were randomly assigned to one of three conditions and completed 13 trials. Participants’ reports accurately reflected the true ratio of the rooms in all conditions; however, participants reported less extreme ratios as the ratio increased. The results suggest that important spatial relationships are preserved in impossible spaces, namely 1) judged sizes of individual rooms, and 2) judgments concerning the relative sizes of different rooms.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In VR, “impossible spaces” allow virtual rooms to occupy overlapping portions of physical space. We investigated how size judgments of impossible spaces are influenced by the ratio of the sizes of overlapping rooms. Participants (n=36) were randomly assigned to one of three conditions and completed 13 trials. Participants’ reports accurately reflected the true ratio of the rooms in all conditions; however, participants reported less extreme ratios as the ratio increased. The results suggest that important spatial relationships are preserved in impossible spaces, namely 1) judged sizes of individual rooms, and 2) judgments concerning the relative sizes of different rooms.", "fno": "09090660", "keywords": [ "Virtual Environments", "Layout", "Conferences", "Legged Locomotion", "Computer Architecture", "Complexity Theory", "Space Exploration", "Human Centered Computing", "Visualization", "Visualization Techniques", "Treemaps", "Human Centered Computing", "Visualization", "Visualization Design And Evaluation Methods This Template Is For Papers Of VGTC Sponsored Conferences Which Are Not Published In A Special Issue Of TVCG" ], "authors": [ { "affiliation": "Clemson University", "fullName": "Catherine Barwulor", "givenName": "Catherine", "surname": "Barwulor", "__typename": "ArticleAuthorType" }, { "affiliation": "Clemson University", "fullName": "Andrew Robb", "givenName": "Andrew", "surname": "Robb", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-03-01T00:00:00", "pubType": "proceedings", "pages": "640-641", "year": "2020", "issn": null, "isbn": "978-1-7281-6532-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09090612", "articleId": "1jIxpbO6bLy", "__typename": "AdjacentArticleType" }, "next": { "fno": "09090461", "articleId": "1jIxzvZw4YU", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2017/6647/0/07892335", "title": "Designing intentional impossible spaces in virtual reality narratives: A case study", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892335/12OmNApcu9b", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2013/4795/0/06549386", "title": "Flexible spaces: A virtual step outside of reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549386/12OmNBOllfZ", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2014/2871/0/06802054", "title": "Time perception during walking in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2014/06802054/12OmNBpmDG4", "parentPublication": { "id": "proceedings/vr/2014/2871/0", "title": "2014 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2013/6097/0/06550194", "title": "Flexible spaces: Dynamic layout generation for infinite walking in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2013/06550194/12OmNyFU75b", "parentPublication": { "id": "proceedings/3dui/2013/6097/0", "title": "2013 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446167", "title": "Redirected Spaces: Going Beyond Borders", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446167/13bd1fph1xv", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040555", "title": "Impossible Spaces: Maximizing Natural Walking in Virtual Environments with Self-Overlapping Architecture", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040555/13rRUygBwhF", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2022/9617/0/961700a379", "title": "Effects of Virtual Room Size and Objects on Relative Translation Gain Thresholds in Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/vr/2022/961700a379/1CJcsYYBYJi", "parentPublication": { "id": "proceedings/vr/2022/9617/0", "title": "2022 IEEE on Conference Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2022/8402/0/840200a622", "title": "Minimaps for Impossible Spaces: Improving Spatial Cognition in Self-Overlapping Virtual Rooms", "doi": null, "abstractUrl": "/proceedings-article/vrw/2022/840200a622/1CJe2Dfxj8Y", "parentPublication": { "id": "proceedings/vrw/2022/8402/0", "title": "2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798095", "title": "Distance Judgments to On- and Off-Ground Objects in Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798095/1cJ0Yxz6rrG", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2021/4057/0/405700a389", "title": "Impossible Open Spaces: Exploring the Effects of Occlusion on the Noticeability of Self-Overlapping Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/vrw/2021/405700a389/1tnWx88Rxuw", "parentPublication": { "id": "proceedings/vrw/2021/4057/0", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1tnWwqMuCzu", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "acronym": "vrw", "groupId": "1836626", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1tnWx88Rxuw", "doi": "10.1109/VRW52623.2021.00081", "title": "Impossible Open Spaces: Exploring the Effects of Occlusion on the Noticeability of Self-Overlapping Virtual Environments", "normalizedTitle": "Impossible Open Spaces: Exploring the Effects of Occlusion on the Noticeability of Self-Overlapping Virtual Environments", "abstract": "Impossible spaces uses self-overlapping virtual architecture to in-crease the area of virtual environments that can be explored by walking. This paper details a preliminary study using psychophysical methods to explore if users' ability to detect overlapping virtual spaces changes when the walls of overlapping rooms and the paths connecting them are removed. The results indicate that impossible spaces can be deployed, even when no walls are presented between adjacent rooms. However, the absence of walls appears to make it easier for users to detect impossible spaces.", "abstracts": [ { "abstractType": "Regular", "content": "Impossible spaces uses self-overlapping virtual architecture to in-crease the area of virtual environments that can be explored by walking. This paper details a preliminary study using psychophysical methods to explore if users' ability to detect overlapping virtual spaces changes when the walls of overlapping rooms and the paths connecting them are removed. The results indicate that impossible spaces can be deployed, even when no walls are presented between adjacent rooms. However, the absence of walls appears to make it easier for users to detect impossible spaces.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Impossible spaces uses self-overlapping virtual architecture to in-crease the area of virtual environments that can be explored by walking. This paper details a preliminary study using psychophysical methods to explore if users' ability to detect overlapping virtual spaces changes when the walls of overlapping rooms and the paths connecting them are removed. The results indicate that impossible spaces can be deployed, even when no walls are presented between adjacent rooms. However, the absence of walls appears to make it easier for users to detect impossible spaces.", "fno": "405700a389", "keywords": [ "Gait Analysis", "Human Computer Interaction", "Virtual Reality", "Impossible Open Spaces", "Virtual Environments", "Self Overlapping Virtual Architecture", "Psychophysical Methods", "Walking", "Occlusion Effects", "Legged Locomotion", "Three Dimensional Displays", "Conferences", "Virtual Environments", "Estimation", "Forestry", "User Interfaces", "Human Centered Computing", "Human Computer Interaction HCI", "Interaction Paradigms", "Virtual Reality" ], "authors": [ { "affiliation": "Aalborg University", "fullName": "Claudiu Ciumedean", "givenName": "Claudiu", "surname": "Ciumedean", "__typename": "ArticleAuthorType" }, { "affiliation": "Aalborg University", "fullName": "Cristian Patras", "givenName": "Cristian", "surname": "Patras", "__typename": "ArticleAuthorType" }, { "affiliation": "Aalborg University", "fullName": "Mantas Cibulskis", "givenName": "Mantas", "surname": "Cibulskis", "__typename": "ArticleAuthorType" }, { "affiliation": "Aalborg University", "fullName": "Norbert Váradi", "givenName": "Norbert", "surname": "Váradi", "__typename": "ArticleAuthorType" }, { "affiliation": "Aalborg University", "fullName": "Niels Christian Nilsson", "givenName": "Niels", "surname": "Christian Nilsson", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-03-01T00:00:00", "pubType": "proceedings", "pages": "389-390", "year": "2021", "issn": null, "isbn": "978-1-6654-4057-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "405700a385", "articleId": "1tnXOcjEGJ2", "__typename": "AdjacentArticleType" }, "next": { "fno": "405700a391", "articleId": "1tnXFgLAfSw", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2017/6647/0/07892335", "title": "Designing intentional impossible spaces in virtual reality narratives: A case study", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892335/12OmNApcu9b", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2013/4795/0/06549386", "title": "Flexible spaces: A virtual step outside of reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549386/12OmNBOllfZ", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2013/6097/0/06550194", "title": "Flexible spaces: Dynamic layout generation for infinite walking in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2013/06550194/12OmNyFU75b", "parentPublication": { "id": "proceedings/3dui/2013/6097/0", "title": "2013 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446167", "title": "Redirected Spaces: Going Beyond Borders", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446167/13bd1fph1xv", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040555", "title": "Impossible Spaces: Maximizing Natural Walking in Virtual Environments with Self-Overlapping Architecture", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040555/13rRUygBwhF", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2022/9617/0/961700a167", "title": "Foldable Spaces: An Overt Redirection Approach for Natural Walking in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2022/961700a167/1CJc5J6RYYM", "parentPublication": { "id": "proceedings/vr/2022/9617/0", "title": "2022 IEEE on Conference Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2022/8402/0/840200a622", "title": "Minimaps for Impossible Spaces: Improving Spatial Cognition in Self-Overlapping Virtual Rooms", "doi": null, "abstractUrl": "/proceedings-article/vrw/2022/840200a622/1CJe2Dfxj8Y", "parentPublication": { "id": "proceedings/vrw/2022/8402/0", "title": "2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2022/5365/0/536500a554", "title": "Short-term Path Prediction for Spontaneous Human Locomotion in Arbitrary Virtual Spaces", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2022/536500a554/1J7WabiAcYE", "parentPublication": { "id": "proceedings/ismar-adjunct/2022/5365/0", "title": "2022 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089554", "title": "Shaking Hands in Virtual Space: Recovery in Redirected Walking for Direct Interaction between Two Users", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089554/1jIxaOIHjaw", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090660", "title": "Relative Room Size Judgments in Impossible Spaces", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090660/1jIxqsUN6ik", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1tuAeQeDJja", "title": "2021 IEEE Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1tuAqY26wzm", "doi": "10.1109/VR50410.2021.00025", "title": "Impossible Staircase: Vertically Real Walking in an Infinite Virtual Tower", "normalizedTitle": "Impossible Staircase: Vertically Real Walking in an Infinite Virtual Tower", "abstract": "We present Impossible Staircase, a real-walking virtual reality system that allows users to climb an infinite virtual tower. Our set-up consists of an one-level scaffold and a lifter. A user climbs up the scaffold by real walking on a stairway while wearing a head-mounted display, and gets reset to the ground level by a lifter imperceptibly. By repeating this process, the user perceives an illusion of climbing an infinite number of levels. Our system achieves the illusion by (1) controlling the movement of the lifter to generate reverse and imperceptible motion, (2) guiding the user through the scaffold with delay mechanisms to reset the lifter in time, and (3) procedural generating overlapping structures to enlarge perceived height of each level. We built a working system and demonstrated it with a 15-min experience. With the working system, we conducted user studies to gain deeper insights into vertical motion simulation and vertical real walking in virtual reality.", "abstracts": [ { "abstractType": "Regular", "content": "We present Impossible Staircase, a real-walking virtual reality system that allows users to climb an infinite virtual tower. Our set-up consists of an one-level scaffold and a lifter. A user climbs up the scaffold by real walking on a stairway while wearing a head-mounted display, and gets reset to the ground level by a lifter imperceptibly. By repeating this process, the user perceives an illusion of climbing an infinite number of levels. Our system achieves the illusion by (1) controlling the movement of the lifter to generate reverse and imperceptible motion, (2) guiding the user through the scaffold with delay mechanisms to reset the lifter in time, and (3) procedural generating overlapping structures to enlarge perceived height of each level. We built a working system and demonstrated it with a 15-min experience. With the working system, we conducted user studies to gain deeper insights into vertical motion simulation and vertical real walking in virtual reality.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We present Impossible Staircase, a real-walking virtual reality system that allows users to climb an infinite virtual tower. Our set-up consists of an one-level scaffold and a lifter. A user climbs up the scaffold by real walking on a stairway while wearing a head-mounted display, and gets reset to the ground level by a lifter imperceptibly. By repeating this process, the user perceives an illusion of climbing an infinite number of levels. Our system achieves the illusion by (1) controlling the movement of the lifter to generate reverse and imperceptible motion, (2) guiding the user through the scaffold with delay mechanisms to reset the lifter in time, and (3) procedural generating overlapping structures to enlarge perceived height of each level. We built a working system and demonstrated it with a 15-min experience. With the working system, we conducted user studies to gain deeper insights into vertical motion simulation and vertical real walking in virtual reality.", "fno": "255600a050", "keywords": [ "Helmet Mounted Displays", "Solid Modelling", "User Experience", "User Interfaces", "Virtual Reality", "Working System", "Vertical Motion Simulation", "Impossible Staircase", "Infinite Virtual Tower", "Real Walking Virtual Reality System", "One Level Scaffold", "Head Mounted Display", "Ground Level", "Illusion", "Lifter", "Imperceptible Motion", "Legged Locomotion", "Solid Modeling", "Three Dimensional Displays", "Head Mounted Displays", "Multimedia Systems", "Poles And Towers", "Virtual Reality", "H 5 1 Information Interfaces And Presentation Multimedia Information Systems Artificial", "Augmented", "Virtual Realities" ], "authors": [ { "affiliation": "National Taiwan University,Department of Computer Science and Information Engineering", "fullName": "Jen-Hao Cheng", "givenName": "Jen-Hao", "surname": "Cheng", "__typename": "ArticleAuthorType" }, { "affiliation": "National Taiwan University,Department of Computer Science and Information Engineering", "fullName": "Yi Chen", "givenName": "Yi", "surname": "Chen", "__typename": "ArticleAuthorType" }, { "affiliation": "National Taiwan University,Department of Computer Science and Information Engineering", "fullName": "Ting-Yi Chang", "givenName": "Ting-Yi", "surname": "Chang", "__typename": "ArticleAuthorType" }, { "affiliation": "National Taiwan University,Department of Computer Science and Information Engineering", "fullName": "Hsu-En Lin", "givenName": "Hsu-En", "surname": "Lin", "__typename": "ArticleAuthorType" }, { "affiliation": "National Taiwan University,Department of Computer Science and Information Engineering", "fullName": "Po-Yao Cosmos Wang", "givenName": "Po-Yao Cosmos", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "National Taiwan University,Department of Computer Science and Information Engineering", "fullName": "Lung-Pan Cheng", "givenName": "Lung-Pan", "surname": "Cheng", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-03-01T00:00:00", "pubType": "proceedings", "pages": "50-56", "year": "2021", "issn": null, "isbn": "978-1-6654-1838-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1tuApVpeQsE", "name": "pvr202118380-09417791s1-mm_255600a050.zip", "size": "203 MB", "location": "https://www.computer.org/csdl/api/v1/extra/pvr202118380-09417791s1-mm_255600a050.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "255600a040", "articleId": "1tuAUKpERUc", "__typename": "AdjacentArticleType" }, "next": { "fno": "255600a057", "articleId": "1tuB2GCo0Ug", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/wsc/2002/7614/2/01166455", "title": "Special purpose simulation modeling of tower cranes", "doi": null, "abstractUrl": "/proceedings-article/wsc/2002/01166455/12OmNAXxXfS", "parentPublication": { "id": "proceedings/wsc/2002/7614/2", "title": "Proceedings of the 2002 Winter Simulation Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hicss/1993/3230/4/00284221", "title": "Globalization and the Tower of Babel", "doi": null, "abstractUrl": "/proceedings-article/hicss/1993/00284221/12OmNvq5jHj", "parentPublication": { "id": "proceedings/hicss/1993/3230/4", "title": "1993 The Twenty-sixth Hawaii International Conference on System Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2010/6846/0/05444717", "title": "The implementation of a novel walking interface within an immersive display", "doi": null, "abstractUrl": "/proceedings-article/3dui/2010/05444717/12OmNx19k1n", "parentPublication": { "id": "proceedings/3dui/2010/6846/0", "title": "2010 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2013/4795/0/06549349", "title": "Visual exploration of the infinite canvas", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549349/12OmNzn38XE", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sive/2018/5713/0/08577177", "title": "Influence of hearing your steps and environmental sounds in VR while walking", "doi": null, "abstractUrl": "/proceedings-article/sive/2018/08577177/17D45XoXP3w", "parentPublication": { "id": "proceedings/sive/2018/5713/0", "title": "2018 IEEE 4th VR Workshop on Sonic Interactions for Virtual Environments (SIVE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tm/2020/12/08807304", "title": "Exploring a Brain-Based Cancelable Biometrics for Smart Headwear: Concept, Implementation, and Evaluation", "doi": null, "abstractUrl": "/journal/tm/2020/12/08807304/1cG62SbIZNu", "parentPublication": { "id": "trans/tm", "title": "IEEE Transactions on Mobile Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089543", "title": "The Space Bender: Supporting Natural Walking via Overt Manipulation of the Virtual Environment", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089543/1jIxb9EbKJq", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089452", "title": "Real and Virtual Environment Mismatching Induces Arousal and Alters Movement Behavior", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089452/1jIxcobDHi0", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2020/8508/0/850800a649", "title": "Comparing World and Screen Coordinate Systems in Optical See-Through Head-Mounted Displays for Text Readability while Walking", "doi": null, "abstractUrl": "/proceedings-article/ismar/2020/850800a649/1pysvKFdazS", "parentPublication": { "id": "proceedings/ismar/2020/8508/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/12/09495106", "title": "A Wheelchair Locomotion Interface in a VR Disability Simulation Reduces Implicit Bias", "doi": null, "abstractUrl": "/journal/tg/2022/12/09495106/1vyjtwGIZkQ", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNywfKyu", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "acronym": "ismar", "groupId": "1000465", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNvA1hoG", "doi": "10.1109/ISMAR.2010.5643602", "title": "Camera pose navigation using Augmented Reality", "normalizedTitle": "Camera pose navigation using Augmented Reality", "abstract": "We propose an Augmented Reality (AR) system that helps users take a picture from a designated pose, such as the position and camera angle of an earlier photo. Repeat photography is frequently used to observe and document changes in an object. Our system uses AR technology to estimate camera poses in real time. When a user takes a photo, the camera pose is saved as a “view bookmark”. To support a user in taking a repeat photo, two simple graphics are rendered in an AR viewer on the camera's screen to guide the user to this bookmarked view. The system then uses image adjustment techniques to create an image based on the user's repeat photo that is even closer to the original.", "abstracts": [ { "abstractType": "Regular", "content": "We propose an Augmented Reality (AR) system that helps users take a picture from a designated pose, such as the position and camera angle of an earlier photo. Repeat photography is frequently used to observe and document changes in an object. Our system uses AR technology to estimate camera poses in real time. When a user takes a photo, the camera pose is saved as a “view bookmark”. To support a user in taking a repeat photo, two simple graphics are rendered in an AR viewer on the camera's screen to guide the user to this bookmarked view. The system then uses image adjustment techniques to create an image based on the user's repeat photo that is even closer to the original.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose an Augmented Reality (AR) system that helps users take a picture from a designated pose, such as the position and camera angle of an earlier photo. Repeat photography is frequently used to observe and document changes in an object. Our system uses AR technology to estimate camera poses in real time. When a user takes a photo, the camera pose is saved as a “view bookmark”. To support a user in taking a repeat photo, two simple graphics are rendered in an AR viewer on the camera's screen to guide the user to this bookmarked view. The system then uses image adjustment techniques to create an image based on the user's repeat photo that is even closer to the original.", "fno": "05643602", "keywords": [ "Augmented Reality", "Image Sensors", "Pose Estimation", "Camera Pose Navigation", "Augmented Reality", "Photography", "Camera Pose Estimation", "Image Adjustment Technique", "Cameras", "Navigation", "Augmented Reality", "Photography", "Three Dimensional Displays", "Solid Modeling", "Augmented Reality", "Repeat Photography", "Rephotography", "Camera Pose Navigation" ], "authors": [ { "affiliation": "FujiXerox Co. Ltd., USA", "fullName": "Jun Shingu", "givenName": "Jun", "surname": "Shingu", "__typename": "ArticleAuthorType" }, { "affiliation": "FX Palo Alto Laboratory Inc., USA", "fullName": "Eleanor Rieffel", "givenName": "Eleanor", "surname": "Rieffel", "__typename": "ArticleAuthorType" }, { "affiliation": "FX Palo Alto Laboratory Inc., USA", "fullName": "Don Kimber", "givenName": "Don", "surname": "Kimber", "__typename": "ArticleAuthorType" }, { "affiliation": "FX Palo Alto Laboratory Inc., USA", "fullName": "Jim Vaughan", "givenName": "Jim", "surname": "Vaughan", "__typename": "ArticleAuthorType" }, { "affiliation": "FX Palo Alto Laboratory Inc., USA", "fullName": "Pernilla Qvarfordt", "givenName": "Pernilla", "surname": "Qvarfordt", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Washington, USA", "fullName": "Kathleen Tuite", "givenName": "Kathleen", "surname": "Tuite", "__typename": "ArticleAuthorType" } ], "idPrefix": "ismar", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-10-01T00:00:00", "pubType": "proceedings", "pages": "271-272", "year": "2010", "issn": null, "isbn": "978-1-4244-9343-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05643601", "articleId": "12OmNxvwoTv", "__typename": "AdjacentArticleType" }, "next": { "fno": "05643603", "articleId": "12OmNzZEApL", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/scc/2016/2628/0/2628a299", "title": "XYZ Indoor Navigation through Augmented Reality: A Research in Progress", "doi": null, "abstractUrl": "/proceedings-article/scc/2016/2628a299/12OmNBigFpq", "parentPublication": { "id": "proceedings/scc/2016/2628/0", "title": "2016 IEEE International Conference on Services Computing (SCC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iwar/1999/0359/0/03590003", "title": "Vision-Based Pose Computation: Robust and Accurate Augmented Reality Tracking", "doi": null, "abstractUrl": "/proceedings-article/iwar/1999/03590003/12OmNBuL1fO", "parentPublication": { "id": "proceedings/iwar/1999/0359/0", "title": "Augmented Reality, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cit-iucc-dasc-picom/2015/0154/0/07363191", "title": "Improving Camera Pose Estimation for Indoor Marker-less Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/cit-iucc-dasc-picom/2015/07363191/12OmNwlqhJz", "parentPublication": { "id": "proceedings/cit-iucc-dasc-picom/2015/0154/0", "title": "2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing (CIT/IUCC/DASC/PICOM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948512", "title": "Diminished reality as challenging extension of mixed and augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948512/12OmNy68EMC", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2011/03/mcg2011030056", "title": "Real-Time Camera Pose Estimation for Wide-Area Augmented Reality Applications", "doi": null, "abstractUrl": "/magazine/cg/2011/03/mcg2011030056/13rRUwwJWBf", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/12/07368948", "title": "Pose Estimation for Augmented Reality: A Hands-On Survey", "doi": null, "abstractUrl": "/journal/tg/2016/12/07368948/13rRUxYINfj", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2018/7592/0/08699207", "title": "Browsing Spatial Photography Using Augmented Models", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2018/08699207/19F1RMZ48XC", "parentPublication": { "id": "proceedings/ismar-adjunct/2018/7592/0", "title": "2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08797821", "title": "Ground Camera Images and UAV 3D Model Registration for Outdoor Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08797821/1cJ0QWVjZ5u", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2019/4765/0/476500a222", "title": "Deep Multi-state Object Pose Estimation for Augmented Reality Assembly", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2019/476500a222/1gysnVT7Ka4", "parentPublication": { "id": "proceedings/ismar-adjunct/2019/4765/0", "title": "2019 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2019/4765/0/476500a128", "title": "Multi-vehicle Cooperative Military Training Simulation System Based on Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2019/476500a128/1gysonjGAqA", "parentPublication": { "id": "proceedings/ismar-adjunct/2019/4765/0", "title": "2019 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwwMf3H", "title": "2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "acronym": "ismarw", "groupId": "1810084", "volume": "0", "displayVolume": "0", "year": "2016", "__typename": "ProceedingType" }, "article": { "id": "12OmNvqmUM8", "doi": "10.1109/ISMAR-Adjunct.2016.0068", "title": "Riverwalk: Incorporating Historical Photographs in Public Outdoor Augmented Reality Experiences", "normalizedTitle": "Riverwalk: Incorporating Historical Photographs in Public Outdoor Augmented Reality Experiences", "abstract": "This paper introduces a user-centered Augmented Reality (AR) approach for publishing 2D media archives as interactive content. We discuss the relevant technical considerations for developing an effective application for public outdoor AR experiences that leverage context-specific elements in a challenging, real-world environment. Specifically, we show how a classical marker-less approach can be combined with mobile sensors and geospatial information in order apply our knowledge of the surroundings to the experience itself. Our contributions provide the enabling technology for Chicago 0,0 Riverwalk, a novel app-based AR experience that superimposes historical imagery onto matching views in downtown Chicago, Illinois along an open, pedestrian waterfront located on the bank of the Chicago River. Historical photographs of sites along the river are superimposed onto buildings, bridges, and other architectural features through image-based AR tracking, providing a striking experience of the city's history as rooted in extant locations along the river.", "abstracts": [ { "abstractType": "Regular", "content": "This paper introduces a user-centered Augmented Reality (AR) approach for publishing 2D media archives as interactive content. We discuss the relevant technical considerations for developing an effective application for public outdoor AR experiences that leverage context-specific elements in a challenging, real-world environment. Specifically, we show how a classical marker-less approach can be combined with mobile sensors and geospatial information in order apply our knowledge of the surroundings to the experience itself. Our contributions provide the enabling technology for Chicago 0,0 Riverwalk, a novel app-based AR experience that superimposes historical imagery onto matching views in downtown Chicago, Illinois along an open, pedestrian waterfront located on the bank of the Chicago River. Historical photographs of sites along the river are superimposed onto buildings, bridges, and other architectural features through image-based AR tracking, providing a striking experience of the city's history as rooted in extant locations along the river.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper introduces a user-centered Augmented Reality (AR) approach for publishing 2D media archives as interactive content. We discuss the relevant technical considerations for developing an effective application for public outdoor AR experiences that leverage context-specific elements in a challenging, real-world environment. Specifically, we show how a classical marker-less approach can be combined with mobile sensors and geospatial information in order apply our knowledge of the surroundings to the experience itself. Our contributions provide the enabling technology for Chicago 0,0 Riverwalk, a novel app-based AR experience that superimposes historical imagery onto matching views in downtown Chicago, Illinois along an open, pedestrian waterfront located on the bank of the Chicago River. Historical photographs of sites along the river are superimposed onto buildings, bridges, and other architectural features through image-based AR tracking, providing a striking experience of the city's history as rooted in extant locations along the river.", "fno": "07836489", "keywords": [ "Augmented Reality", "History", "Image Processing", "Multimedia Computing", "User Interfaces", "Historical Photographs", "Public Outdoor Augmented Reality Experiences", "User Centered Augmented Reality", "2 D Media Archive", "Interactive Content", "Context Specific Elements", "Mobile Sensors", "Geospatial Information", "Chicago Riverwalk", "Historical Imagery", "Chicago Illinois", "Pedestrian Waterfront", "Chicago River", "Image Based AR Tracking", "Multimedia Information Systems", "Cameras", "Sensors", "Two Dimensional Displays", "Rivers", "Three Dimensional Displays", "Augmented Reality", "Mobile Communication", "H 5 1 Information Interfaces And Presentation E G HCI Multimedia Information Systems Artificial", "Augmented And Virtual Realities" ], "authors": [ { "affiliation": null, "fullName": "Marco Cavallo", "givenName": "Marco", "surname": "Cavallo", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Geoffrey Alan Rhodes", "givenName": "Geoffrey Alan", "surname": "Rhodes", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Angus Graeme Forbes", "givenName": "Angus Graeme", "surname": "Forbes", "__typename": "ArticleAuthorType" } ], "idPrefix": "ismarw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2016-09-01T00:00:00", "pubType": "proceedings", "pages": "160-165", "year": "2016", "issn": null, "isbn": "978-1-5090-3740-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07836488", "articleId": "12OmNBVrji6", "__typename": "AdjacentArticleType" }, "next": { "fno": "07836490", "articleId": "12OmNx76TJ8", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ismar-amh/2010/9339/0/05643296", "title": "An integrated design flow in user interface and interaction for enhancing mobile AR gaming experiences", "doi": null, "abstractUrl": "/proceedings-article/ismar-amh/2010/05643296/12OmNBE7Moa", "parentPublication": { "id": "proceedings/ismar-amh/2010/9339/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality - Arts, Media, and Humanities", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismarw/2016/3740/0/07836522", "title": "Exploring Immersive AR Instructions for Procedural Tasks: The Role of Depth, Motion, and Volumetric Representations", "doi": null, "abstractUrl": "/proceedings-article/ismarw/2016/07836522/12OmNxETajV", "parentPublication": { "id": "proceedings/ismarw/2016/3740/0", "title": "2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2017/6327/0/6327a176", "title": "[POSTER] Augmented Reality Assistance in the Central Field-of-View Outperforms Peripheral Displays for Order Picking: Results from a Virtual Reality Simulation Study", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2017/6327a176/12OmNxGAKUL", "parentPublication": { "id": "proceedings/ismar-adjunct/2017/6327/0", "title": "2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2018/7592/0/08699264", "title": "RealityMedia: An Experimental Digital Book in WebXR", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2018/08699264/19F1TRfLKda", "parentPublication": { "id": "proceedings/ismar-adjunct/2018/7592/0", "title": "2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/t4e/2019/4227/0/422700a193", "title": "Collaborative Approaches to Problem-Solving on Lines and Angles Using Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/t4e/2019/422700a193/1hgtGoEU0OA", "parentPublication": { "id": "proceedings/t4e/2019/4227/0", "title": "2019 IEEE Tenth International Conference on Technology for Education (T4E)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/inciscos/2019/5581/0/558100a116", "title": "ARTOUR: Augmented Reality for Tourism - A Case Study in Riobamba, Ecuador", "doi": null, "abstractUrl": "/proceedings-article/inciscos/2019/558100a116/1iHUGrjAlXO", "parentPublication": { "id": "proceedings/inciscos/2019/5581/0", "title": "2019 International Conference on Information Systems and Computer Science (INCISCOS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090401", "title": "Learning to Match 2D Images and 3D LiDAR Point Clouds for Outdoor Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090401/1jIxmhXvH7a", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090631", "title": "Framing the Scene: An Examination of Augmented Reality Head Worn Displays in Construction Assembly Tasks", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090631/1jIxyGx0KXK", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2020/8508/0/850800a704", "title": "HydrogenAR: Interactive Data-Driven Presentation of Dispenser Reliability", "doi": null, "abstractUrl": "/proceedings-article/ismar/2020/850800a704/1pyswX2oQBW", "parentPublication": { "id": "proceedings/ismar/2020/8508/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2020/8508/0/850800a064", "title": "The Effects of Object Shape, Fidelity, Color, and Luminance on Depth Perception in Handheld Mobile Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2020/850800a064/1pysxPMqyTm", "parentPublication": { "id": "proceedings/ismar/2020/8508/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAYoKmw", "title": "2013 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "acronym": "ismar", "groupId": "1000465", "volume": "0", "displayVolume": "0", "year": "2013", "__typename": "ProceedingType" }, "article": { "id": "12OmNxGSm5G", "doi": "10.1109/ISMAR.2013.6671790", "title": "Content first - A concept for industrial augmented reality maintenance applications using mobile devices", "normalizedTitle": "Content first - A concept for industrial augmented reality maintenance applications using mobile devices", "abstract": "Although AR has a long history in the area of maintenance and service-support in industry, there still is a lack of lightweight, yet practical solutions for handheld AR systems in everyday workflows. Attempts to support complex maintenance tasks with AR still miss reliable tracking techniques, simple ways to be integrated into existing maintenance environments, and practical authoring solutions, which minimize costs for specialized content generation. We present a general, customisable application framework, allowing to employ AR and VR techniques in order to support technicians in their daily tasks. In contrast to other systems, we do not aim to replace existing support systems such as traditional manuals. Instead we integrate well-known AR- and novel presentation techniques with existing instruction media. To this end practical authoring solutions are crucial and hence we present an application development system based on web-standards such as HTML,CSS and X3D.", "abstracts": [ { "abstractType": "Regular", "content": "Although AR has a long history in the area of maintenance and service-support in industry, there still is a lack of lightweight, yet practical solutions for handheld AR systems in everyday workflows. Attempts to support complex maintenance tasks with AR still miss reliable tracking techniques, simple ways to be integrated into existing maintenance environments, and practical authoring solutions, which minimize costs for specialized content generation. We present a general, customisable application framework, allowing to employ AR and VR techniques in order to support technicians in their daily tasks. In contrast to other systems, we do not aim to replace existing support systems such as traditional manuals. Instead we integrate well-known AR- and novel presentation techniques with existing instruction media. To this end practical authoring solutions are crucial and hence we present an application development system based on web-standards such as HTML,CSS and X3D.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Although AR has a long history in the area of maintenance and service-support in industry, there still is a lack of lightweight, yet practical solutions for handheld AR systems in everyday workflows. Attempts to support complex maintenance tasks with AR still miss reliable tracking techniques, simple ways to be integrated into existing maintenance environments, and practical authoring solutions, which minimize costs for specialized content generation. We present a general, customisable application framework, allowing to employ AR and VR techniques in order to support technicians in their daily tasks. In contrast to other systems, we do not aim to replace existing support systems such as traditional manuals. Instead we integrate well-known AR- and novel presentation techniques with existing instruction media. To this end practical authoring solutions are crucial and hence we present an application development system based on web-standards such as HTML,CSS and X3D.", "fno": "06671790", "keywords": [ "Three Dimensional Displays", "Maintenance Engineering", "Solid Modeling", "Augmented Reality", "Context", "Media", "Documentation" ], "authors": [ { "affiliation": "Fraunhofer IGD, Darmstadt, Germany", "fullName": "Timo Engelke", "givenName": "Timo", "surname": "Engelke", "__typename": "ArticleAuthorType" }, { "affiliation": "Fraunhofer IGD, Darmstadt, Germany", "fullName": "Jens Keil", "givenName": "Jens", "surname": "Keil", "__typename": "ArticleAuthorType" }, { "affiliation": "Fraunhofer IGD, Darmstadt, Germany", "fullName": "Pavel Rojtberg", "givenName": "Pavel", "surname": "Rojtberg", "__typename": "ArticleAuthorType" }, { "affiliation": "Fraunhofer IGD, Darmstadt, Germany", "fullName": "Folker Wientapper", "givenName": "Folker", "surname": "Wientapper", "__typename": "ArticleAuthorType" }, { "affiliation": "Fraunhofer IGD, Darmstadt, Germany", "fullName": "Sabine Webel", "givenName": "Sabine", "surname": "Webel", "__typename": "ArticleAuthorType" }, { "affiliation": "Fraunhofer IGD, Darmstadt, Germany", "fullName": "Ulrich Bockholt", "givenName": "Ulrich", "surname": "Bockholt", "__typename": "ArticleAuthorType" } ], "idPrefix": "ismar", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2013-10-01T00:00:00", "pubType": "proceedings", "pages": "251-252", "year": "2013", "issn": null, "isbn": "978-1-4799-2869-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06671789", "articleId": "12OmNvmG7Xx", "__typename": "AdjacentArticleType" }, "next": { "fno": "06671791", "articleId": "12OmNyRg4vp", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ismar-adjunct/2017/6327/0/6327a039", "title": "[POSTER] Industrial Augmented Reality: Transferring a Numerical Control Connected Augmented Realty System from Marketing to Maintenance", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2017/6327a039/12OmNzC5Tr7", "parentPublication": { "id": "proceedings/ismar-adjunct/2017/6327/0", "title": "2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ec/2014/04/07004838", "title": "Challenges, Opportunities, and Future Trends of Emerging Techniques for Augmented Reality-Based Maintenance", "doi": null, "abstractUrl": "/journal/ec/2014/04/07004838/13rRUwbs27k", "parentPublication": { "id": "trans/ec", "title": "IEEE Transactions on Emerging Topics in Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2018/7592/0/08699183", "title": "Industrial Augmented Reality: Requirements for an Augmented Reality Maintenance Worker Support System", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2018/08699183/19F1MWRWSqs", "parentPublication": { "id": "proceedings/ismar-adjunct/2018/7592/0", "title": "2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2022/5365/0/536500a678", "title": "Industrial Augmented Reality: Lessons learned from a long-term On-site Assessment of Augmented Reality Maintenance Worker Support Systems", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2022/536500a678/1J7WeihLonS", "parentPublication": { "id": "proceedings/ismar-adjunct/2022/5365/0", "title": "2022 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2022/5365/0/536500a931", "title": "From Lab to Reality: Optimization of Industrial Augmented Reality Interfaces", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2022/536500a931/1J7Wo6Bhxa8", "parentPublication": { "id": "proceedings/ismar-adjunct/2022/5365/0", "title": "2022 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2020/7675/0/767500a067", "title": "Industrial Augmented Reality: Concepts and User Interface Designs for Augmented Reality Maintenance Worker Support Systems", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2020/767500a067/1pBMhXqBhCM", "parentPublication": { "id": "proceedings/ismar-adjunct/2020/7675/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2020/7675/0/767500a203", "title": "Industrial Augmented Reality: 3D-Content Editor for Augmented Reality Maintenance Worker Support System", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2020/767500a203/1pBMigKK7F6", "parentPublication": { "id": "proceedings/ismar-adjunct/2020/7675/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2020/7675/0/767500a172", "title": "Design preferences on Industrial Augmented Reality: a survey with potential technical writers", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2020/767500a172/1pBMjARVuEg", "parentPublication": { "id": "proceedings/ismar-adjunct/2020/7675/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2021/1298/0/129800a268", "title": "Mobile Augmented Reality as a Field-Assistance Tool in Urban Maintenance", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2021/129800a268/1yeQL1J7Xsk", "parentPublication": { "id": "proceedings/ismar-adjunct/2021/1298/0", "title": "2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2021/1298/0/129800a212", "title": "A Nugget-Based Concept for Creating Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2021/129800a212/1yfxJj6swzC", "parentPublication": { "id": "proceedings/ismar-adjunct/2021/1298/0", "title": "2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNrNh0vs", "title": "2013 23rd International Conference on Artificial Reality and Telexistence (ICAT)", "acronym": "icat", "groupId": "1001485", "volume": "0", "displayVolume": "0", "year": "2013", "__typename": "ProceedingType" }, "article": { "id": "12OmNxiKrZY", "doi": "10.1109/ICAT.2013.6728899", "title": "Reality and perception: Utilization of many facets of augmented reality", "normalizedTitle": "Reality and perception: Utilization of many facets of augmented reality", "abstract": "What we perceive with our senses becomes the basic real time information that directs and helps with the tasks we do. Our perception may represent the real world environment or what is perceived is a view of the real world modified with simulated elements of the environment. Augmented reality (AR) represents a system where a view of a live real physical environment is supplemented by computer-generated elements such as sound, video, graphic or location data. Development of AR systems has been facilitated by the advancement in both hardware and software technologies, making it easier to implement all functionalities in real time. Such real time enhancements have been helped by new techniques in such areas as computer vision, object recognition, and registration methodologies which enable user to interact with and manipulate the real world environment effectively. AR advanced computer graphics is integrated to real world data and viewed by the user using see-through head mounted display (HMD). This immersion of the integrated real and simulated world provides a unique perception to users. What he sees, feels, hears, and smells is indistinguishable between what is real and what is computer-generated. Technology has propelled AR systems from experimental laboratories to the marketplace demonstrating great promise in the fields of arts, architecture, archaeology, commerce, construction, education, entertainment, gaming, system maintenance, medical and military applications.", "abstracts": [ { "abstractType": "Regular", "content": "What we perceive with our senses becomes the basic real time information that directs and helps with the tasks we do. Our perception may represent the real world environment or what is perceived is a view of the real world modified with simulated elements of the environment. Augmented reality (AR) represents a system where a view of a live real physical environment is supplemented by computer-generated elements such as sound, video, graphic or location data. Development of AR systems has been facilitated by the advancement in both hardware and software technologies, making it easier to implement all functionalities in real time. Such real time enhancements have been helped by new techniques in such areas as computer vision, object recognition, and registration methodologies which enable user to interact with and manipulate the real world environment effectively. AR advanced computer graphics is integrated to real world data and viewed by the user using see-through head mounted display (HMD). This immersion of the integrated real and simulated world provides a unique perception to users. What he sees, feels, hears, and smells is indistinguishable between what is real and what is computer-generated. Technology has propelled AR systems from experimental laboratories to the marketplace demonstrating great promise in the fields of arts, architecture, archaeology, commerce, construction, education, entertainment, gaming, system maintenance, medical and military applications.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "What we perceive with our senses becomes the basic real time information that directs and helps with the tasks we do. Our perception may represent the real world environment or what is perceived is a view of the real world modified with simulated elements of the environment. Augmented reality (AR) represents a system where a view of a live real physical environment is supplemented by computer-generated elements such as sound, video, graphic or location data. Development of AR systems has been facilitated by the advancement in both hardware and software technologies, making it easier to implement all functionalities in real time. Such real time enhancements have been helped by new techniques in such areas as computer vision, object recognition, and registration methodologies which enable user to interact with and manipulate the real world environment effectively. AR advanced computer graphics is integrated to real world data and viewed by the user using see-through head mounted display (HMD). This immersion of the integrated real and simulated world provides a unique perception to users. What he sees, feels, hears, and smells is indistinguishable between what is real and what is computer-generated. Technology has propelled AR systems from experimental laboratories to the marketplace demonstrating great promise in the fields of arts, architecture, archaeology, commerce, construction, education, entertainment, gaming, system maintenance, medical and military applications.", "fno": "06728899", "keywords": [ "Glass", "Three Dimensional Displays", "Augmented Reality", "Solid Modeling", "Real Time Systems", "Computational Modeling", "Google", "Displays", "Virtual Reality", "Augmented Reality", "Simulated Reality" ], "authors": [ { "affiliation": "Adjunct Fac., Univ. of Houston, Houston, TX, USA", "fullName": "Zafar Taqvi", "givenName": "Zafar", "surname": "Taqvi", "__typename": "ArticleAuthorType" } ], "idPrefix": "icat", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2013-12-01T00:00:00", "pubType": "proceedings", "pages": "11-12", "year": "2013", "issn": null, "isbn": "978-4-904490-11-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06728898", "articleId": "12OmNAq3hRD", "__typename": "AdjacentArticleType" }, "next": { "fno": "06728900", "articleId": "12OmNzmclxt", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2008/1971/0/04480794", "title": "The Effects of Virtual Reality, Augmented Reality, and Motion Parallax on Egocentric Depth Perception", "doi": null, "abstractUrl": "/proceedings-article/vr/2008/04480794/12OmNrAMERg", "parentPublication": { "id": "proceedings/vr/2008/1971/0", "title": "IEEE Virtual Reality 2008", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2009/3890/0/3890a187", "title": "Dual Face Interaction in Handheld Augmented Reality Environments", "doi": null, "abstractUrl": "/proceedings-article/ism/2009/3890a187/12OmNxGj9VX", "parentPublication": { "id": "proceedings/ism/2009/3890/0", "title": "2009 11th IEEE International Symposium on Multimedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismarw/2016/3740/0/07836540", "title": "InspectAR: An Augmented Reality Inspection Framework for Industry", "doi": null, "abstractUrl": "/proceedings-article/ismarw/2016/07836540/12OmNyfdOIx", "parentPublication": { "id": "proceedings/ismarw/2016/3740/0", "title": "2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvvrhc/1998/8283/0/82830078", "title": "Vision and Graphics in Producing Mixed Reality Worlds", "doi": null, "abstractUrl": "/proceedings-article/cvvrhc/1998/82830078/12OmNylbov1", "parentPublication": { "id": "proceedings/cvvrhc/1998/8283/0", "title": "Computer Vision for Virtual Reality Based Human Communications, Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismarw/2015/8471/0/8471a012", "title": "Depth Perception and Action in Wearable Augmented Reality: A Pilot Study", "doi": null, "abstractUrl": "/proceedings-article/ismarw/2015/8471a012/12OmNzd7bXH", "parentPublication": { "id": "proceedings/ismarw/2015/8471/0", "title": "2015 IEEE International Symposium on Mixed and Augmented Reality Workshops (ISMARW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446175", "title": "A User-Based Comparison of Two Augmented Reality Glasses", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446175/13bd1eTtWYo", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08797992", "title": "Text Presentation for Augmented Reality Applications in Dual-Task Situations", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08797992/1cJ0SIvkZnG", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceee/2019/3910/0/391000a079", "title": "Desktop Artillery Simulation Using Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/iceee/2019/391000a079/1cpqGEpXo5O", "parentPublication": { "id": "proceedings/iceee/2019/3910/0", "title": "2019 6th International Conference on Electrical and Electronics Engineering (ICEEE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090604", "title": "Utilizing AR Glasses as Mobility Aid for People with Low Vision", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090604/1jIxi1ubEcg", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/12/09199574", "title": "Gaze-Dependent Simulation of Light Perception in Virtual Reality", "doi": null, "abstractUrl": "/journal/tg/2020/12/09199574/1ncgnMqzLJm", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwMXnv0", "title": "2014 IEEE Virtual Reality (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNyRPgKH", "doi": "10.1109/VR.2014.6802045", "title": "Transitional Augmented Reality navigation for live captured scenes", "normalizedTitle": "Transitional Augmented Reality navigation for live captured scenes", "abstract": "Augmented Reality (AR) applications require knowledge about the real world environment in which they are used. This knowledge is often gathered while developing the AR application and stored for future uses of the application. Consequently, changes to the real world lead to a mismatch between the previously recorded data and the real world. New capturing techniques based on dense Simultaneous Localization and Mapping (SLAM) not only allow users to capture real world scenes at run-time, but also enables them to capture changes of the world. However, instead of using previously recorded and prepared scenes, users must interact with an unprepared environment. In this paper, we present a set of new interaction techniques that support users in handling captured real world environments. The techniques present virtual viewpoints of the scene based on a scene analysis and provide natural transitions between the AR view and virtual viewpoints. We demonstrate our approach with a SLAM based prototype that allows us to capture a real world scene and describe example applications of our system.", "abstracts": [ { "abstractType": "Regular", "content": "Augmented Reality (AR) applications require knowledge about the real world environment in which they are used. This knowledge is often gathered while developing the AR application and stored for future uses of the application. Consequently, changes to the real world lead to a mismatch between the previously recorded data and the real world. New capturing techniques based on dense Simultaneous Localization and Mapping (SLAM) not only allow users to capture real world scenes at run-time, but also enables them to capture changes of the world. However, instead of using previously recorded and prepared scenes, users must interact with an unprepared environment. In this paper, we present a set of new interaction techniques that support users in handling captured real world environments. The techniques present virtual viewpoints of the scene based on a scene analysis and provide natural transitions between the AR view and virtual viewpoints. We demonstrate our approach with a SLAM based prototype that allows us to capture a real world scene and describe example applications of our system.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Augmented Reality (AR) applications require knowledge about the real world environment in which they are used. This knowledge is often gathered while developing the AR application and stored for future uses of the application. Consequently, changes to the real world lead to a mismatch between the previously recorded data and the real world. New capturing techniques based on dense Simultaneous Localization and Mapping (SLAM) not only allow users to capture real world scenes at run-time, but also enables them to capture changes of the world. However, instead of using previously recorded and prepared scenes, users must interact with an unprepared environment. In this paper, we present a set of new interaction techniques that support users in handling captured real world environments. The techniques present virtual viewpoints of the scene based on a scene analysis and provide natural transitions between the AR view and virtual viewpoints. We demonstrate our approach with a SLAM based prototype that allows us to capture a real world scene and describe example applications of our system.", "fno": "06802045", "keywords": [ "Cameras", "Switches", "Navigation", "Image Reconstruction", "Geometry", "Semantics", "Simultaneous Localization And Mapping", "H 5 2 Information Interfaces And Presentation User Interfaces Interaction Styles", "H 5 1 Information Interfaces And Presentation Multimedia Information Systems Artificial Augmented And Virtual Realities" ], "authors": [ { "affiliation": "Graz University of Technology", "fullName": "Markus Tatzgern", "givenName": "Markus", "surname": "Tatzgern", "__typename": "ArticleAuthorType" }, { "affiliation": "Graz University of Technology", "fullName": "Raphael Grasset", "givenName": "Raphael", "surname": "Grasset", "__typename": "ArticleAuthorType" }, { "affiliation": "Graz University of Technology", "fullName": "Denis Kalkofen", "givenName": "Denis", "surname": "Kalkofen", "__typename": "ArticleAuthorType" }, { "affiliation": "Graz University of Technology", "fullName": "Dieter Schmalstieg", "givenName": "Dieter", "surname": "Schmalstieg", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-03-01T00:00:00", "pubType": "proceedings", "pages": "21-26", "year": "2014", "issn": null, "isbn": "978-1-4799-2871-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06802044", "articleId": "12OmNz4SOCN", "__typename": "AdjacentArticleType" }, "next": { "fno": "06802046", "articleId": "12OmNzTppDb", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dv/2017/2610/0/261001a356", "title": "Loop-Closure Detection in Urban Scenes for Autonomous Robot Navigation", "doi": null, "abstractUrl": "/proceedings-article/3dv/2017/261001a356/12OmNAndiqV", "parentPublication": { "id": "proceedings/3dv/2017/2610/0", "title": "2017 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2017/6327/0/6327a135", "title": "[POSTER] Semantic Augmented Reality Environment with Material-Aware Physical Interactions", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2017/6327a135/12OmNB6UIa5", "parentPublication": { "id": "proceedings/ismar-adjunct/2017/6327/0", "title": "2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2015/8332/0/8332a607", "title": "Reconstructing Street-Scenes in Real-Time from a Driving Car", "doi": null, "abstractUrl": "/proceedings-article/3dv/2015/8332a607/12OmNvm6VJS", "parentPublication": { "id": "proceedings/3dv/2015/8332/0", "title": "2015 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948422", "title": "Dense planar SLAM", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948422/12OmNx7G661", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2018/8425/0/842500a616", "title": "Multi-planar Monocular Reconstruction of Manhattan Indoor Scenes", "doi": null, "abstractUrl": "/proceedings-article/3dv/2018/842500a616/17D45XvMcbo", "parentPublication": { "id": "proceedings/3dv/2018/8425/0", "title": "2018 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500b858", "title": "SIGNAV: Semantically-Informed GPS-Denied Navigation and Mapping in Visually-Degraded Environments", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500b858/1B13SA7R4M8", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600m2776", "title": "NICE-SLAM: Neural Implicit Scalable Encoding for SLAM", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600m2776/1H1jhZSaE0M", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798315", "title": "A Hybrid RTK GNSS and SLAM Outdoor Augmented Reality System", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798315/1cJ0Soon8Yg", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2020/9134/0/913400a708", "title": "Guido: Augmented Reality for Indoor Navigation Using Commodity Hardware", "doi": null, "abstractUrl": "/proceedings-article/iv/2020/913400a708/1rSR9YRSTpS", "parentPublication": { "id": "proceedings/iv/2020/9134/0", "title": "2020 24th International Conference Information Visualisation (IV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900c814", "title": "Differentiable SLAM-net: Learning Particle SLAM for Visual Navigation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900c814/1yeIHPaF1WU", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyQYteG", "title": "Ninth IEEE International Symposium on Wearable Computers (ISWC'05)", "acronym": "iswc", "groupId": "1000810", "volume": "0", "displayVolume": "0", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "12OmNzVoBzm", "doi": "10.1109/ISWC.2005.41", "title": "Pictorial Depth Cues for Outdoor Augmented Reality", "normalizedTitle": "Pictorial Depth Cues for Outdoor Augmented Reality", "abstract": "This paper presents and evaluates a set of pictorial depth cues for far-field outdoor mobile augmented reality (AR). We examine the problem of accurately placing virtual annotations at physical target points from a static point of view. While it is easy to line up annotations with a target point's projection in the view plane, finding the correct distance for the annotation is difficult if the target point is not represented in an environment model. We have found that AR depth cues, such as vertical and horizontal shadow planes, a small top-down map, or color encodings of relative depth, have a positive impact on a user's ability to align a 3D cursor with physical objects at various distances. These cues aid the user's depth perception and estimation by providing information about the 3D cursor's distance and its relationship in 3- space to any features that may already have been annotated. We conducted a user study that measures the effects of different depth cues for both absolute 3D cursor placement as well as placement relative to a small number of marked reference points, whose distances are known. Our study provides insight about mobile AR users' ability to judge distances both absolutely and relatively, and we identify techniques that successfully enhance their performance.", "abstracts": [ { "abstractType": "Regular", "content": "This paper presents and evaluates a set of pictorial depth cues for far-field outdoor mobile augmented reality (AR). We examine the problem of accurately placing virtual annotations at physical target points from a static point of view. While it is easy to line up annotations with a target point's projection in the view plane, finding the correct distance for the annotation is difficult if the target point is not represented in an environment model. We have found that AR depth cues, such as vertical and horizontal shadow planes, a small top-down map, or color encodings of relative depth, have a positive impact on a user's ability to align a 3D cursor with physical objects at various distances. These cues aid the user's depth perception and estimation by providing information about the 3D cursor's distance and its relationship in 3- space to any features that may already have been annotated. We conducted a user study that measures the effects of different depth cues for both absolute 3D cursor placement as well as placement relative to a small number of marked reference points, whose distances are known. Our study provides insight about mobile AR users' ability to judge distances both absolutely and relatively, and we identify techniques that successfully enhance their performance.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper presents and evaluates a set of pictorial depth cues for far-field outdoor mobile augmented reality (AR). We examine the problem of accurately placing virtual annotations at physical target points from a static point of view. While it is easy to line up annotations with a target point's projection in the view plane, finding the correct distance for the annotation is difficult if the target point is not represented in an environment model. We have found that AR depth cues, such as vertical and horizontal shadow planes, a small top-down map, or color encodings of relative depth, have a positive impact on a user's ability to align a 3D cursor with physical objects at various distances. These cues aid the user's depth perception and estimation by providing information about the 3D cursor's distance and its relationship in 3- space to any features that may already have been annotated. We conducted a user study that measures the effects of different depth cues for both absolute 3D cursor placement as well as placement relative to a small number of marked reference points, whose distances are known. Our study provides insight about mobile AR users' ability to judge distances both absolutely and relatively, and we identify techniques that successfully enhance their performance.", "fno": "24190092", "keywords": [], "authors": [ { "affiliation": "University of California, Santa Barbara", "fullName": "Jason Wither", "givenName": "Jason", "surname": "Wither", "__typename": "ArticleAuthorType" }, { "affiliation": "University of California, Santa Barbara", "fullName": "Tobias H?llerer", "givenName": "Tobias", "surname": "H?llerer", "__typename": "ArticleAuthorType" } ], "idPrefix": "iswc", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2005-10-01T00:00:00", "pubType": "proceedings", "pages": "92-99", "year": "2005", "issn": null, "isbn": "0-7695-2419-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "24190086", "articleId": "12OmNBRKwC3", "__typename": "AdjacentArticleType" }, "next": { "fno": "24190100", "articleId": "12OmNBuL1i7", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cgiv/2010/4166/0/4166a173", "title": "Perceptually-Guided Design of Nonperspectives through Pictorial Depth Cues", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2010/4166a173/12OmNAtaS3V", "parentPublication": { "id": "proceedings/cgiv/2010/4166/0", "title": "2010 Seventh International Conference on Computer Graphics, Imaging and Visualization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2007/1630/0/04408900", "title": "Novel Depth Cues from Uncalibrated Near-field Lighting", "doi": null, "abstractUrl": "/proceedings-article/iccv/2007/04408900/12OmNy7Qfl3", "parentPublication": { "id": "proceedings/iccv/2007/1630/0", "title": "2007 11th IEEE International Conference on Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2009/3943/0/04810999", "title": "Indoor vs. Outdoor Depth Perception for Mobile Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2009/04810999/12OmNzvhvIR", "parentPublication": { "id": "proceedings/vr/2009/3943/0", "title": "2009 IEEE Virtual Reality Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446580", "title": "An Investigation of Head Motion and Perceptual Motion Cues' Influence on User Depth Perception of Augmented Reality Neurosurgical Simulators", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446580/13bd1eTtWYv", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/03/v0429", "title": "Egocentric Depth Judgments in Optical, See-Through Augmented Reality", "doi": null, "abstractUrl": "/journal/tg/2007/03/v0429/13rRUxYrbM5", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/11/07164348", "title": "Matching and Reaching Depth Judgments with Real and Augmented Reality Targets", "doi": null, "abstractUrl": "/journal/tg/2015/11/07164348/13rRUxZzAhH", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/02/08462792", "title": "The Effect of Focal Distance, Age, and Brightness on Near-Field Augmented Reality Depth Matching", "doi": null, "abstractUrl": "/journal/tg/2020/02/08462792/13w3loWnQPK", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798095", "title": "Distance Judgments to On- and Off-Ground Objects in Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798095/1cJ0Yxz6rrG", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/05/08998348", "title": "Effects of Depth Information on Visual Target Identification Task Performance in Shared Gaze Environments", "doi": null, "abstractUrl": "/journal/tg/2020/05/08998348/1hrXedrZXos", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2021/1838/0/255600a122", "title": "Augmented Reality for Maritime Navigation Assistance - Egocentric Depth Perception in Large Distance Outdoor Environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2021/255600a122/1tuB9Rs5D2M", "parentPublication": { "id": "proceedings/vr/2021/1838/0", "title": "2021 IEEE Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1MNgk3BHlS0", "title": "2023 IEEE Conference Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2023", "__typename": "ProceedingType" }, "article": { "id": "1MNgLMRvOtq", "doi": "10.1109/VR55154.2023.00064", "title": "Cross-View Visual Geo-Localization for Outdoor Augmented Reality", "normalizedTitle": "Cross-View Visual Geo-Localization for Outdoor Augmented Reality", "abstract": "Precise estimation of global orientation and location is critical to ensure a compelling outdoor Augmented Reality (AR) experience. We address the problem of geo-pose estimation by cross-view matching of query ground images to a geo-referenced aerial satellite image database. Recently, neural network-based methods have shown state-of-the-art performance in cross-view matching. However, most of the prior works focus only on location estimation, ignoring orientation, which cannot meet the requirements in outdoor AR applications. We propose a new transformer neural network-based model and a modified triplet ranking loss for joint location and orientation estimation. Experiments on several benchmark cross-view geo-localization datasets show that our model achieves state-of-the-art performance. Furthermore, we present an approach to extend the single image query-based geo-localization approach by utilizing temporal information from a navigation pipeline for robust continuous geo-localization. Experimentation on several large-scale real-world video sequences demonstrates that our approach enables high-precision and stable AR insertion.", "abstracts": [ { "abstractType": "Regular", "content": "Precise estimation of global orientation and location is critical to ensure a compelling outdoor Augmented Reality (AR) experience. We address the problem of geo-pose estimation by cross-view matching of query ground images to a geo-referenced aerial satellite image database. Recently, neural network-based methods have shown state-of-the-art performance in cross-view matching. However, most of the prior works focus only on location estimation, ignoring orientation, which cannot meet the requirements in outdoor AR applications. We propose a new transformer neural network-based model and a modified triplet ranking loss for joint location and orientation estimation. Experiments on several benchmark cross-view geo-localization datasets show that our model achieves state-of-the-art performance. Furthermore, we present an approach to extend the single image query-based geo-localization approach by utilizing temporal information from a navigation pipeline for robust continuous geo-localization. Experimentation on several large-scale real-world video sequences demonstrates that our approach enables high-precision and stable AR insertion.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Precise estimation of global orientation and location is critical to ensure a compelling outdoor Augmented Reality (AR) experience. We address the problem of geo-pose estimation by cross-view matching of query ground images to a geo-referenced aerial satellite image database. Recently, neural network-based methods have shown state-of-the-art performance in cross-view matching. However, most of the prior works focus only on location estimation, ignoring orientation, which cannot meet the requirements in outdoor AR applications. We propose a new transformer neural network-based model and a modified triplet ranking loss for joint location and orientation estimation. Experiments on several benchmark cross-view geo-localization datasets show that our model achieves state-of-the-art performance. Furthermore, we present an approach to extend the single image query-based geo-localization approach by utilizing temporal information from a navigation pipeline for robust continuous geo-localization. Experimentation on several large-scale real-world video sequences demonstrates that our approach enables high-precision and stable AR insertion.", "fno": "481500a493", "keywords": [ "Visualization", "Solid Modeling", "Three Dimensional Displays", "Satellites", "Video Sequences", "Estimation", "Benchmark Testing", "Cross View Visual Geo Localization", "Ground Aerial Matching", "Outdoor Augmented Reality", "Transformer Neural Network", "Visual Inertial Navigation" ], "authors": [ { "affiliation": "Center for Vision Technologies, SRI International,Princeton,NJ", "fullName": "Niluthpol Chowdhury Mithun", "givenName": "Niluthpol Chowdhury", "surname": "Mithun", "__typename": "ArticleAuthorType" }, { "affiliation": "Center for Vision Technologies, SRI International,Princeton,NJ", "fullName": "Kshitij S. Minhas", "givenName": "Kshitij S.", "surname": "Minhas", "__typename": "ArticleAuthorType" }, { "affiliation": "Center for Vision Technologies, SRI International,Princeton,NJ", "fullName": "Han-Pang Chiu", "givenName": "Han-Pang", "surname": "Chiu", "__typename": "ArticleAuthorType" }, { "affiliation": "Center for Vision Technologies, SRI International,Princeton,NJ", "fullName": "Taragay Oskiper", "givenName": "Taragay", "surname": "Oskiper", "__typename": "ArticleAuthorType" }, { "affiliation": "Center for Vision Technologies, SRI International,Princeton,NJ", "fullName": "Mikhail Sizintsev", "givenName": "Mikhail", "surname": "Sizintsev", "__typename": "ArticleAuthorType" }, { "affiliation": "Center for Vision Technologies, SRI International,Princeton,NJ", "fullName": "Supun Samarasekera", "givenName": "Supun", "surname": "Samarasekera", "__typename": "ArticleAuthorType" }, { "affiliation": "Center for Vision Technologies, SRI International,Princeton,NJ", "fullName": "Rakesh Kumar", "givenName": "Rakesh", "surname": "Kumar", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2023-03-01T00:00:00", "pubType": "proceedings", "pages": "493-502", "year": "2023", "issn": null, "isbn": "979-8-3503-4815-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1MNgLBJKyoE", "name": "pvr202348150-010108434s1-mm_481500a493.zip", "size": "74.5 MB", "location": "https://www.computer.org/csdl/api/v1/extra/pvr202348150-010108434s1-mm_481500a493.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "481500a482", "articleId": "1MNgRYfCXU4", "__typename": "AdjacentArticleType" }, "next": { "fno": "481500a503", "articleId": "1MNgIE9xnBC", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "trans/tg/2015/11/07164332", "title": "Instant Outdoor Localization and SLAM Initialization from 2.5D Maps", "doi": null, "abstractUrl": "/journal/tg/2015/11/07164332/13rRUxBa5c1", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000i601", "title": "Benchmarking 6DOF Outdoor Visual Localization in Changing Conditions", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000i601/17D45W2Wyzk", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000h258", "title": "CVM-Net: Cross-View Matching Network for Image-Based Ground-to-Aerial Geo-Localization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000h258/17D45W9KVJX", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2023/03/09826417", "title": "Accurate 3-DoF Camera Geo-Localization via Ground-to-Satellite Image Matching", "doi": null, "abstractUrl": "/journal/tp/2023/03/09826417/1EVdA0iYdFu", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600c913", "title": "Cross-View Image Sequence Geo-localization", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600c913/1L8qqTqQeli", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2019/9552/0/955200a742", "title": "GEOCAPSNET: Ground to Aerial View Image Geo-Localization using Capsule Network", "doi": null, "abstractUrl": "/proceedings-article/icme/2019/955200a742/1cdOFZKk2S4", "parentPublication": { "id": "proceedings/icme/2019/9552/0", "title": "2019 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300f617", "title": "Lending Orientation to Neural Networks for Cross-View Geo-Localization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300f617/1gyrRdxyCWY", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2019/4765/0/476500a037", "title": "A High-Precision Localization Device for Outdoor Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2019/476500a037/1gyskTjecsE", "parentPublication": { "id": "proceedings/ismar-adjunct/2019/4765/0", "title": "2019 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900d639", "title": "VIGOR: Cross-View Image Geo-localization beyond One-to-one Retrieval", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900d639/1yeKLsUXtHW", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900g484", "title": "Coming Down to Earth: Satellite-to-Street View Synthesis for Geo-Localization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900g484/1yeLgyCSbEQ", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1cI6akLvAuQ", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1cJ0QWVjZ5u", "doi": "10.1109/VR.2019.8797821", "title": "Ground Camera Images and UAV 3D Model Registration for Outdoor Augmented Reality", "normalizedTitle": "Ground Camera Images and UAV 3D Model Registration for Outdoor Augmented Reality", "abstract": "This paper presents a novel virtual-real registration approach for augmented reality (AR) in large-scale outdoor environments. Essentially, it is a pose estimation for the mobile camera images (ground camera images) in 3D model recovered by Unmanned Aerial Vehicle (UAV) image sequence via Structure-From-Motion (SFM) technology. The approach considers to indirectly establish the spatial relationship between 2D and 3D space by inferring the transformation relationship between the ground camera images and the UAV 3D model rendered images. Specifically, the proposed approach can overcome the positioning errors, which are deterioration and drift in the GPS, and deviation of orientation. The experimental results demonstrate the possibility of the proposed virtual-real registration approach, and show that the approach is robust, efficient and intuitive for AR in large-scale outdoor environments.", "abstracts": [ { "abstractType": "Regular", "content": "This paper presents a novel virtual-real registration approach for augmented reality (AR) in large-scale outdoor environments. Essentially, it is a pose estimation for the mobile camera images (ground camera images) in 3D model recovered by Unmanned Aerial Vehicle (UAV) image sequence via Structure-From-Motion (SFM) technology. The approach considers to indirectly establish the spatial relationship between 2D and 3D space by inferring the transformation relationship between the ground camera images and the UAV 3D model rendered images. Specifically, the proposed approach can overcome the positioning errors, which are deterioration and drift in the GPS, and deviation of orientation. The experimental results demonstrate the possibility of the proposed virtual-real registration approach, and show that the approach is robust, efficient and intuitive for AR in large-scale outdoor environments.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper presents a novel virtual-real registration approach for augmented reality (AR) in large-scale outdoor environments. Essentially, it is a pose estimation for the mobile camera images (ground camera images) in 3D model recovered by Unmanned Aerial Vehicle (UAV) image sequence via Structure-From-Motion (SFM) technology. The approach considers to indirectly establish the spatial relationship between 2D and 3D space by inferring the transformation relationship between the ground camera images and the UAV 3D model rendered images. Specifically, the proposed approach can overcome the positioning errors, which are deterioration and drift in the GPS, and deviation of orientation. The experimental results demonstrate the possibility of the proposed virtual-real registration approach, and show that the approach is robust, efficient and intuitive for AR in large-scale outdoor environments.", "fno": "08797821", "keywords": [ "Augmented Reality", "Autonomous Aerial Vehicles", "Cameras", "Control Engineering Computing", "Image Registration", "Image Sequences", "Pose Estimation", "Remotely Operated Vehicles", "Rendering Computer Graphics", "Solid Modelling", "Unmanned Aerial Vehicle Image Sequence", "Virtual Real Registration Approach", "Structure From Motion Technology", "SFM Technology", "Mobile Camera Images", "Outdoor Augmented Reality", "UAV 3 D Model Registration", "Ground Camera Images", "Large Scale Outdoor Environments", "Solid Modeling", "Three Dimensional Displays", "Cameras", "Unmanned Aerial Vehicles", "Mobile Handsets", "Global Positioning System", "Augmented Reality", "Virtual Real Registration", "Outdoor AR", "Cross Domain Image Matching", "Human Centered Computing", "Visualization", "Visualization Techniques", "Augmented Reality" ], "authors": [ { "affiliation": "Fujian Key Laboratory of Sensing and Computing for Smart City, School of Information Science and Engineering, Xiamen University, Xiamen, China", "fullName": "Weiquan Liu", "givenName": "Weiquan", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "Fujian Key Laboratory of Sensing and Computing for Smart City, School of Information Science and Engineering, Xiamen University, Xiamen, China", "fullName": "Cheng Wang", "givenName": "Cheng", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "Fujian Key Laboratory of Sensing and Computing for Smart City, School of Information Science and Engineering, Xiamen University, Xiamen, China", "fullName": "Yu Zang", "givenName": "Yu", "surname": "Zang", "__typename": "ArticleAuthorType" }, { "affiliation": "National Tsing Hua University, Taiwan", "fullName": "Shang-Hong Lai", "givenName": "Shang-Hong", "surname": "Lai", "__typename": "ArticleAuthorType" }, { "affiliation": "Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, China", "fullName": "Dongdong Weng", "givenName": "Dongdong", "surname": "Weng", "__typename": "ArticleAuthorType" }, { "affiliation": "Fujian Key Laboratory of Sensing and Computing for Smart City, School of Information Science and Engineering, Xiamen University, Xiamen, China", "fullName": "Xuesheng Bian", "givenName": "Xuesheng", "surname": "Bian", "__typename": "ArticleAuthorType" }, { "affiliation": "Fujian Key Laboratory of Sensing and Computing for Smart City, School of Information Science and Engineering, Xiamen University, Xiamen, China", "fullName": "Xiuhong Lin", "givenName": "Xiuhong", "surname": "Lin", "__typename": "ArticleAuthorType" }, { "affiliation": "Fujian Key Laboratory of Sensing and Computing for Smart City, School of Information Science and Engineering, Xiamen University, Xiamen, China", "fullName": "Xuelun Shen", "givenName": "Xuelun", "surname": "Shen", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Waterloo, Waterloo, Canada", "fullName": "Jonathan Li", "givenName": "Jonathan", "surname": "Li", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-03-01T00:00:00", "pubType": "proceedings", "pages": "1050-1051", "year": "2019", "issn": null, "isbn": "978-1-7281-1377-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08798224", "articleId": "1cJ1aEpf4wo", "__typename": "AdjacentArticleType" }, "next": { "fno": "08797782", "articleId": "1cJ153S8Qb6", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/irc/2018/4652/0/465201a266", "title": "Autonomous Navigation and Sensorless Obstacle Avoidance for UGV with Environment Information from UAV", "doi": null, "abstractUrl": "/proceedings-article/irc/2018/465201a266/12OmNwpoFJH", "parentPublication": { "id": "proceedings/irc/2018/4652/0", "title": "2018 Second IEEE International Conference on Robotic Computing (IRC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscid/2008/3311/2/3311b069", "title": "Development Actuality and Application of Registration Technology in Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/iscid/2008/3311b069/12OmNxFsmHr", "parentPublication": { "id": "proceedings/iscid/2008/3311/2", "title": "2008 International Symposium on Computational Intelligence and Design", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cybersecsym/2016/5771/0/07942421", "title": "Security Testing of an Unmanned Aerial Vehicle (UAV)", "doi": null, "abstractUrl": "/proceedings-article/cybersecsym/2016/07942421/12OmNxuXcB0", "parentPublication": { "id": "proceedings/cybersecsym/2016/5771/0", "title": "2016 Cybersecurity Symposium (CYBERSEC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2018/6100/0/610000b594", "title": "A Deep CNN-Based Framework For Enhanced Aerial Imagery Registration with Applications to UAV Geolocalization", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2018/610000b594/17D45WgziNJ", "parentPublication": { "id": "proceedings/cvprw/2018/6100/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmcce/2019/4689/0/468900a578", "title": "Design of Civil UAV Counter System Based on BDS", "doi": null, "abstractUrl": "/proceedings-article/icmcce/2019/468900a578/1h0FhqOGU3m", "parentPublication": { "id": "proceedings/icmcce/2019/4689/0", "title": "2019 4th International Conference on Mechanical, Control and Computer Engineering (ICMCCE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tm/2022/07/09263339", "title": "<monospace>RFEye</monospace> in the Sky", "doi": null, "abstractUrl": "/journal/tm/2022/07/09263339/1oReJNWNVwA", "parentPublication": { "id": "trans/tm", "title": "IEEE Transactions on Mobile Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2020/7675/0/767500a104", "title": "User-Aided Global Registration Method using Geospatial 3D Data for Large-Scale Mobile Outdoor Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2020/767500a104/1pBMkF3iMYU", "parentPublication": { "id": "proceedings/ismar-adjunct/2020/7675/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ipccc/2020/9829/0/09391527", "title": "LIDAUS: Localization of IoT Device via Anchor UAV SLAM", "doi": null, "abstractUrl": "/proceedings-article/ipccc/2020/09391527/1syuAMNDGz6", "parentPublication": { "id": "proceedings/ipccc/2020/9829/0", "title": "2020 IEEE 39th International Performance Computing and Communications Conference (IPCCC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ec/2022/03/09519550", "title": "Fast and Robust UAV to UAV Detection and Tracking From Video", "doi": null, "abstractUrl": "/journal/ec/2022/03/09519550/1wc8Vbe1r7G", "parentPublication": { "id": "trans/ec", "title": "IEEE Transactions on Emerging Topics in Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icwcsg/2021/2598/0/259800a029", "title": "Research on Key Technologies Of UAV Navigation and Positioning System", "doi": null, "abstractUrl": "/proceedings-article/icwcsg/2021/259800a029/1yQBeTbXnji", "parentPublication": { "id": "proceedings/icwcsg/2021/2598/0", "title": "2021 International Conference on Wireless Communications and Smart Grid (ICWCSG)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1pBMeBWXAZ2", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "acronym": "ismar-adjunct", "groupId": "1810084", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1pBMkF3iMYU", "doi": "10.1109/ISMAR-Adjunct51615.2020.00041", "title": "User-Aided Global Registration Method using Geospatial 3D Data for Large-Scale Mobile Outdoor Augmented Reality", "normalizedTitle": "User-Aided Global Registration Method using Geospatial 3D Data for Large-Scale Mobile Outdoor Augmented Reality", "abstract": "Accurate global camera registration is a key requirement for precise AR visualizations in large-scale outdoor AR applications. Existing approaches mostly use complex image-based registration methods requiring large pre-registered databases of geo-referenced images or point clouds that are hardly applicable to large-scale areas. In this paper, we present a simple yet effective user-aided registration method that utilizes common geospatial 3D data to globally register mobile devices. For this purpose, text-based 3D geospatial data including digital 3D terrain and city models is processed into small-scale 3D meshes and displayed in a live AR view. Via two common mobile touch gestures the generated virtual models can be aligned manually to match the actual perception of the real-world environment. Experimental results show that - combined with a robust local visual-inertial tracking system - this approach enables an efficient and accurate global registration of mobile devices in various environments determining the camera attitude with less than one degree deviation while achieving a high degree of immersion through realistic occlusion behavior.", "abstracts": [ { "abstractType": "Regular", "content": "Accurate global camera registration is a key requirement for precise AR visualizations in large-scale outdoor AR applications. Existing approaches mostly use complex image-based registration methods requiring large pre-registered databases of geo-referenced images or point clouds that are hardly applicable to large-scale areas. In this paper, we present a simple yet effective user-aided registration method that utilizes common geospatial 3D data to globally register mobile devices. For this purpose, text-based 3D geospatial data including digital 3D terrain and city models is processed into small-scale 3D meshes and displayed in a live AR view. Via two common mobile touch gestures the generated virtual models can be aligned manually to match the actual perception of the real-world environment. Experimental results show that - combined with a robust local visual-inertial tracking system - this approach enables an efficient and accurate global registration of mobile devices in various environments determining the camera attitude with less than one degree deviation while achieving a high degree of immersion through realistic occlusion behavior.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Accurate global camera registration is a key requirement for precise AR visualizations in large-scale outdoor AR applications. Existing approaches mostly use complex image-based registration methods requiring large pre-registered databases of geo-referenced images or point clouds that are hardly applicable to large-scale areas. In this paper, we present a simple yet effective user-aided registration method that utilizes common geospatial 3D data to globally register mobile devices. For this purpose, text-based 3D geospatial data including digital 3D terrain and city models is processed into small-scale 3D meshes and displayed in a live AR view. Via two common mobile touch gestures the generated virtual models can be aligned manually to match the actual perception of the real-world environment. Experimental results show that - combined with a robust local visual-inertial tracking system - this approach enables an efficient and accurate global registration of mobile devices in various environments determining the camera attitude with less than one degree deviation while achieving a high degree of immersion through realistic occlusion behavior.", "fno": "767500a104", "keywords": [ "Augmented Reality", "Cameras", "Data Visualisation", "Geophysical Image Processing", "Image Registration", "Mobile Computing", "Solid Modelling", "AR Visualizations", "Large Scale Outdoor AR Applications", "Geo Referenced Images", "Mobile Devices", "Text Based 3 D Geospatial Data", "City Models", "Small Scale 3 D Meshes", "Visual Inertial Tracking System", "Large Scale Mobile Outdoor Augmented Reality", "User Aided Global Registration", "Geospatial 3 D Data", "User Aided Registration", "Global Camera Registration", "Digital 3 D Terrain", "Live AR View", "Mobile Touch Gestures", "Camera Attitude", "Realistic Occlusion Behavior", "Solid Modeling", "Three Dimensional Displays", "Urban Areas", "Cameras", "Mobile Handsets", "Geospatial Analysis", "Augmented Reality", "Computing Methodologies", "Computer Graphics", "Graphics Systems And Interfaces", "Mixed Augmented Reality", "Human Centered Computing", "Human Computer Interaction HCI", "Interaction Paradigms", "Mixed Augmented Reality" ], "authors": [ { "affiliation": "University of Applied Science,HTW Berlin,Germany", "fullName": "Simon Burkard", "givenName": "Simon", "surname": "Burkard", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Applied Science,HTW Berlin,Germany", "fullName": "Frank Fuchs-Kittowski", "givenName": "Frank", "surname": "Fuchs-Kittowski", "__typename": "ArticleAuthorType" } ], "idPrefix": "ismar-adjunct", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-11-01T00:00:00", "pubType": "proceedings", "pages": "104-109", "year": "2020", "issn": null, "isbn": "978-1-7281-7675-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "767500a102", "articleId": "1pBMeQmOrAc", "__typename": "AdjacentArticleType" }, "next": { "fno": "767500a110", "articleId": "1pBMgrzCp8c", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icnc/2008/3304/6/3304f138", "title": "Cognition-Based Augmented Reality Visualization of the Geospatial Data", "doi": null, "abstractUrl": "/proceedings-article/icnc/2008/3304f138/12OmNqESuhG", "parentPublication": { "id": "proceedings/icnc/2008/3304/6", "title": "2008 Fourth International Conference on Natural Computation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismarw/2016/3740/0/07836489", "title": "Riverwalk: Incorporating Historical Photographs in Public Outdoor Augmented Reality Experiences", "doi": null, "abstractUrl": "/proceedings-article/ismarw/2016/07836489/12OmNvqmUM8", "parentPublication": { "id": "proceedings/ismarw/2016/3740/0", "title": "2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2006/01/mcg2006010014", "title": "3D Modeling with the Tinmith Mobile Outdoor Augmented Reality System", "doi": null, "abstractUrl": "/magazine/cg/2006/01/mcg2006010014/13rRUB6Sq2O", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08797821", "title": "Ground Camera Images and UAV 3D Model Registration for Outdoor Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08797821/1cJ0QWVjZ5u", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090401", "title": "Learning to Match 2D Images and 3D LiDAR Point Clouds for Outdoor Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090401/1jIxmhXvH7a", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2020/1485/0/09106030", "title": "Simplifying the Process of Creating Augmented Outdoor Scenes", "doi": null, "abstractUrl": "/proceedings-article/icmew/2020/09106030/1kwqK8F3v6o", "parentPublication": { "id": "proceedings/icmew/2020/1485/0", "title": "2020 IEEE International Conference on Multimedia & Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iisa/2020/2346/0/09284355", "title": "Augmented Reality applications in Cultural Heritage: Technology or Content?", "doi": null, "abstractUrl": "/proceedings-article/iisa/2020/09284355/1pttMNwrFMQ", "parentPublication": { "id": "proceedings/iisa/2020/2346/0", "title": "2020 11th International Conference on Information, Intelligence, Systems and Applications (IISA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icris/2020/1969/0/196900a422", "title": "Integrated analysis of urban multiple geospatial information based on GIS", "doi": null, "abstractUrl": "/proceedings-article/icris/2020/196900a422/1wG60oyQvFm", "parentPublication": { "id": "proceedings/icris/2020/1969/0", "title": "2020 International Conference on Robots & Intelligent System (ICRIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2021/1298/0/129800a367", "title": "A Classification of Augmented Reality Approaches for Spatial Data Visualization", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2021/129800a367/1yeQHlFhK0w", "parentPublication": { "id": "proceedings/ismar-adjunct/2021/1298/0", "title": "2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscc/2021/2744/0/09631438", "title": "Mobile Augmented Reality for Craniotomy Planning", "doi": null, "abstractUrl": "/proceedings-article/iscc/2021/09631438/1zmvEvuTSCI", "parentPublication": { "id": "proceedings/iscc/2021/2744/0", "title": "2021 IEEE Symposium on Computers and Communications (ISCC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNynsbxl", "title": "2014 2nd International Conference on 3D Vision (3DV)", "acronym": "3dv", "groupId": "1800494", "volume": "1", "displayVolume": "1", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNBdJ5j1", "doi": "10.1109/3DV.2014.112", "title": "Lighting Estimation in Outdoor Image Collections", "normalizedTitle": "Lighting Estimation in Outdoor Image Collections", "abstract": "Large scale structure-from-motion (SfM) algorithms have recently enabled the reconstruction of highly detailed 3-D models of our surroundings simply by taking photographs. In this paper, we propose to leverage these reconstruction techniques to automatically estimate the outdoor illumination conditions for each image in a SfM photo collection. We introduce a novel dataset of outdoor photo collections, where the ground truth lighting conditions are known at each image. We also present an inverse rendering approach that recovers a high dynamic range estimate of the lighting conditions for each low dynamic range input image. Our novel database is used to quantitatively evaluate the performance of our algorithm. Results show that physically plausible lighting estimates can faithfully be recovered, both in terms of light direction and intensity.", "abstracts": [ { "abstractType": "Regular", "content": "Large scale structure-from-motion (SfM) algorithms have recently enabled the reconstruction of highly detailed 3-D models of our surroundings simply by taking photographs. In this paper, we propose to leverage these reconstruction techniques to automatically estimate the outdoor illumination conditions for each image in a SfM photo collection. We introduce a novel dataset of outdoor photo collections, where the ground truth lighting conditions are known at each image. We also present an inverse rendering approach that recovers a high dynamic range estimate of the lighting conditions for each low dynamic range input image. Our novel database is used to quantitatively evaluate the performance of our algorithm. Results show that physically plausible lighting estimates can faithfully be recovered, both in terms of light direction and intensity.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Large scale structure-from-motion (SfM) algorithms have recently enabled the reconstruction of highly detailed 3-D models of our surroundings simply by taking photographs. In this paper, we propose to leverage these reconstruction techniques to automatically estimate the outdoor illumination conditions for each image in a SfM photo collection. We introduce a novel dataset of outdoor photo collections, where the ground truth lighting conditions are known at each image. We also present an inverse rendering approach that recovers a high dynamic range estimate of the lighting conditions for each low dynamic range input image. Our novel database is used to quantitatively evaluate the performance of our algorithm. Results show that physically plausible lighting estimates can faithfully be recovered, both in terms of light direction and intensity.", "fno": "7000a131", "keywords": [ "Lighting", "Sun", "Probes", "Databases", "Cameras", "Estimation", "Optimization", "High Dynamic Range", "Lighting Estimation", "3 D Reconstruction" ], "authors": [ { "affiliation": "Laval Univ., Quebec City, QC, Canada", "fullName": "Jean-Francois Lalonde", "givenName": "Jean-Francois", "surname": "Lalonde", "__typename": "ArticleAuthorType" }, { "affiliation": "Disney Res., Pittsburgh, PA, USA", "fullName": "Iain Matthews", "givenName": "Iain", "surname": "Matthews", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-12-01T00:00:00", "pubType": "proceedings", "pages": "131-138", "year": "2014", "issn": null, "isbn": "978-1-4799-7000-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "7000a123", "articleId": "12OmNzwpUnX", "__typename": "AdjacentArticleType" }, "next": { "fno": "7000a139", "articleId": "12OmNAXxXid", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2017/1032/0/1032e529", "title": "Learning High Dynamic Range from Outdoor Panoramas", "doi": null, "abstractUrl": "/proceedings-article/iccv/2017/1032e529/12OmNwHz09M", "parentPublication": { "id": "proceedings/iccv/2017/1032/0", "title": "2017 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2017/0457/0/0457c373", "title": "Deep outdoor illumination estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457c373/12OmNy1SFMf", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2013/2246/0/2246a132", "title": "Estimation of Environmental Lighting from Known Geometries for Mobile Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/cw/2013/2246a132/12OmNy5R3C7", "parentPublication": { "id": "proceedings/cw/2013/2246/0", "title": "2013 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2015/8332/0/8332a460", "title": "Shadow Detection and Sun Direction in Photo Collections", "doi": null, "abstractUrl": "/proceedings-article/3dv/2015/8332a460/12OmNzwpUq0", "parentPublication": { "id": "proceedings/3dv/2015/8332/0", "title": "2015 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2023/04/09678000", "title": "Real-Time Lighting Estimation for Augmented Reality via Differentiable Screen-Space Rendering", "doi": null, "abstractUrl": "/journal/tg/2023/04/09678000/1A4SuYWCI7K", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300g920", "title": "Deep Sky Modeling for Single Image Outdoor Lighting Estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300g920/1gyrdbEY2sM", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300k0150", "title": "All-Weather Deep Outdoor Lighting Estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300k0150/1gyrg6Ricuc", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/05/08998303", "title": "Physically-inspired Deep Light Estimation from a Homogeneous-Material Object for Mixed Reality Lighting", "doi": null, "abstractUrl": "/journal/tg/2020/05/08998303/1hrXfo1lGb6", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a355", "title": "High-Dynamic-Range Lighting Estimation From Face Portraits", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a355/1qyxlbQeCtO", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900m2829", "title": "Spatially-Varying Outdoor Lighting Estimation from Intrinsics", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900m2829/1yeJsK0NeLK", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBSBk5H", "title": "Computer-Aided Design and Computer Graphics, International Conference on", "acronym": "cadgraphics", "groupId": "1001488", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNqyUUzK", "doi": "10.1109/CAD/Graphics.2011.51", "title": "On-line Illumination Estimation of Outdoor Scenes Based on Area Selection for Augmented Reality", "normalizedTitle": "On-line Illumination Estimation of Outdoor Scenes Based on Area Selection for Augmented Reality", "abstract": "In augmented reality, consistent illumination plays an important role when integrating a virtual object into a video of real scene. In this paper, we propose a novel image based framework to estimateon-line the dynamic ally changing illumination parameters of outdoor video sequences captured by a fixed camera. Unlike previous approaches which either request to know the scene geometry or involve huge storage to preserve time dependent basis images or statistic parameters, our approach requires very simple interaction at the initialization stage by a few brushes to select areas with specified surface normal, which are used to calculate the sunlight parameters. An optimization procedure is also applied, ensuring the robustness and precision of our estimation. Experimental results demonstrate the effectiveness and flexibility of the proposed approach.", "abstracts": [ { "abstractType": "Regular", "content": "In augmented reality, consistent illumination plays an important role when integrating a virtual object into a video of real scene. In this paper, we propose a novel image based framework to estimateon-line the dynamic ally changing illumination parameters of outdoor video sequences captured by a fixed camera. Unlike previous approaches which either request to know the scene geometry or involve huge storage to preserve time dependent basis images or statistic parameters, our approach requires very simple interaction at the initialization stage by a few brushes to select areas with specified surface normal, which are used to calculate the sunlight parameters. An optimization procedure is also applied, ensuring the robustness and precision of our estimation. Experimental results demonstrate the effectiveness and flexibility of the proposed approach.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In augmented reality, consistent illumination plays an important role when integrating a virtual object into a video of real scene. In this paper, we propose a novel image based framework to estimateon-line the dynamic ally changing illumination parameters of outdoor video sequences captured by a fixed camera. Unlike previous approaches which either request to know the scene geometry or involve huge storage to preserve time dependent basis images or statistic parameters, our approach requires very simple interaction at the initialization stage by a few brushes to select areas with specified surface normal, which are used to calculate the sunlight parameters. An optimization procedure is also applied, ensuring the robustness and precision of our estimation. Experimental results demonstrate the effectiveness and flexibility of the proposed approach.", "fno": "4497a439", "keywords": [ "Augmented Reality", "Outdoor Scenes", "Illumination Estimation", "Area Selection" ], "authors": [ { "affiliation": null, "fullName": "Guanyu Xing", "givenName": "Guanyu", "surname": "Xing", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yanli Liu", "givenName": "Yanli", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Xueying Qin", "givenName": "Xueying", "surname": "Qin", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Qunsheng Peng", "givenName": "Qunsheng", "surname": "Peng", "__typename": "ArticleAuthorType" } ], "idPrefix": "cadgraphics", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-09-01T00:00:00", "pubType": "proceedings", "pages": "439-442", "year": "2011", "issn": null, "isbn": "978-0-7695-4497-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4497a433", "articleId": "12OmNyTfg3h", "__typename": "AdjacentArticleType" }, "next": { "fno": "4497a443", "articleId": "12OmNAfy7Id", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icfn/2010/3940/0/3940a016", "title": "An Effective Background Subtraction under a Continuosly and Rapidly Varying Illumination", "doi": null, "abstractUrl": "/proceedings-article/icfn/2010/3940a016/12OmNBCZnSu", "parentPublication": { "id": "proceedings/icfn/2010/3940/0", "title": "Future Networks, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icfn/2010/3940/0/3940a024", "title": "Background Subtraction under Single Varying Illumination in Different Background Luminance", "doi": null, "abstractUrl": "/proceedings-article/icfn/2010/3940a024/12OmNBf94Xj", "parentPublication": { "id": "proceedings/icfn/2010/3940/0", "title": "Future Networks, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2012/4660/0/06402547", "title": "Instant indirect illumination for dynamic mixed reality scenes", "doi": null, "abstractUrl": "/proceedings-article/ismar/2012/06402547/12OmNxWcH5T", "parentPublication": { "id": "proceedings/ismar/2012/4660/0", "title": "2012 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2017/0457/0/0457c373", "title": "Deep outdoor illumination estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457c373/12OmNy1SFMf", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscv/1995/7190/0/71900467", "title": "Direct recovery of motion and range from images of scenes with time-varying illumination", "doi": null, "abstractUrl": "/proceedings-article/iscv/1995/71900467/12OmNzBOi4x", "parentPublication": { "id": "proceedings/iscv/1995/7190/0", "title": "Computer Vision, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/1995/7042/0/70420720", "title": "Color constancy under varying illumination", "doi": null, "abstractUrl": "/proceedings-article/iccv/1995/70420720/12OmNzlUKQX", "parentPublication": { "id": "proceedings/iccv/1995/7042/0", "title": "Computer Vision, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2013/02/ttg2013020210", "title": "Rich Intrinsic Image Decomposition of Outdoor Scenes from Multiple Views", "doi": null, "abstractUrl": "/journal/tg/2013/02/ttg2013020210/13rRUILtJm9", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2000/04/v0289", "title": "Interactive Virtual Relighting of Real Scenes", "doi": null, "abstractUrl": "/journal/tg/2000/04/v0289/13rRUwbaqUD", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040573", "title": "Online Tracking of Outdoor Lighting Variations for Augmented Reality with Moving Cameras", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040573/13rRUyY28Yr", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2003/03/i0290", "title": "Illumination from Shadows", "doi": null, "abstractUrl": "/journal/tp/2003/03/i0290/13rRUygT7z3", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwdbV00", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNrY3LEd", "doi": "10.1109/CVPR.2012.6247684", "title": "Photometric stereo for outdoor webcams", "normalizedTitle": "Photometric stereo for outdoor webcams", "abstract": "We present a photometric stereo technique that operates on time-lapse sequences captured by static outdoor webcams over the course of several months. Outdoor webcams produce a large set of uncontrolled images subject to varying lighting and weather conditions. We first automatically select a suitable subset of the captured frames for further processing, reducing the dataset size by several orders of magnitude. A camera calibration step is applied to recover the camera response function, the absolute camera orientation, and to compute the light directions for each image. Finally, we describe a new photometric stereo technique for non-Lambertian scenes and unknown light source intensities to recover normal maps and spatially varying materials of the scene.", "abstracts": [ { "abstractType": "Regular", "content": "We present a photometric stereo technique that operates on time-lapse sequences captured by static outdoor webcams over the course of several months. Outdoor webcams produce a large set of uncontrolled images subject to varying lighting and weather conditions. We first automatically select a suitable subset of the captured frames for further processing, reducing the dataset size by several orders of magnitude. A camera calibration step is applied to recover the camera response function, the absolute camera orientation, and to compute the light directions for each image. Finally, we describe a new photometric stereo technique for non-Lambertian scenes and unknown light source intensities to recover normal maps and spatially varying materials of the scene.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We present a photometric stereo technique that operates on time-lapse sequences captured by static outdoor webcams over the course of several months. Outdoor webcams produce a large set of uncontrolled images subject to varying lighting and weather conditions. We first automatically select a suitable subset of the captured frames for further processing, reducing the dataset size by several orders of magnitude. A camera calibration step is applied to recover the camera response function, the absolute camera orientation, and to compute the light directions for each image. Finally, we describe a new photometric stereo technique for non-Lambertian scenes and unknown light source intensities to recover normal maps and spatially varying materials of the scene.", "fno": "034P1A34", "keywords": [ "Stereo Image Processing", "Calibration", "Cameras", "Photometric Light Sources", "Photometry", "Spatially Varying Materials", "Photometric Stereo Technique", "Time Lapse Sequences", "Static Outdoor Webcams", "Uncontrolled Images", "Varying Lighting Conditions", "Weather Conditions", "Captured Frames", "Dataset Size", "Camera Calibration Step", "Camera Response Function", "Absolute Camera Orientation", "Light Directions", "Non Lambertian Scenes", "Light Source Intensity", "Normal Maps", "Cameras", "Materials", "Sun", "Lighting", "Calibration", "Light Sources", "Image Reconstruction" ], "authors": [ { "affiliation": "Tech. Univ. Darmstadt, Darmstadt, Germany", "fullName": "F. Langguth", "givenName": "F.", "surname": "Langguth", "__typename": "ArticleAuthorType" }, { "affiliation": "Tech. Univ. Darmstadt, Darmstadt, Germany", "fullName": "J. Ackermann", "givenName": "J.", "surname": "Ackermann", "__typename": "ArticleAuthorType" }, { "affiliation": "Tech. Univ. Darmstadt, Darmstadt, Germany", "fullName": "S. Fuhrmann", "givenName": "S.", "surname": "Fuhrmann", "__typename": "ArticleAuthorType" }, { "affiliation": "Tech. Univ. Darmstadt, Darmstadt, Germany", "fullName": "M. Goesele", "givenName": "M.", "surname": "Goesele", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-06-01T00:00:00", "pubType": "proceedings", "pages": "262-269", "year": "2012", "issn": "1063-6919", "isbn": "978-1-4673-1226-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "033P1A33", "articleId": "12OmNzYNN7C", "__typename": "AdjacentArticleType" }, "next": { "fno": "035P1A35", "articleId": "12OmNBSSVos", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccp/2015/8667/0/07168379", "title": "What Is a Good Day for Outdoor Photometric Stereo?", "doi": null, "abstractUrl": "/proceedings-article/iccp/2015/07168379/12OmNBmf3by", "parentPublication": { "id": "proceedings/iccp/2015/8667/0", "title": "2015 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2015/8332/0/8332a028", "title": "x-Hour Outdoor Photometric Stereo", "doi": null, "abstractUrl": "/proceedings-article/3dv/2015/8332a028/12OmNCesr5K", "parentPublication": { "id": "proceedings/3dv/2015/8332/0", "title": "2015 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2013/6463/0/06528306", "title": "Outdoor photometric stereo", "doi": null, "abstractUrl": "/proceedings-article/iccp/2013/06528306/12OmNrNh0Dh", "parentPublication": { "id": "proceedings/iccp/2013/6463/0", "title": "2013 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2011/707/0/05753116", "title": "Single-shot photometric stereo by spectral multiplexing", "doi": null, "abstractUrl": "/proceedings-article/iccp/2011/05753116/12OmNvkGW9o", "parentPublication": { "id": "proceedings/iccp/2011/707/0", "title": "IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d478", "title": "Photometric Stereo with Small Angular Variations", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d478/12OmNwE9OqB", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2015/6683/0/6683a302", "title": "Photometric Stereo in the Wild", "doi": null, "abstractUrl": "/proceedings-article/wacv/2015/6683a302/12OmNznkK53", "parentPublication": { "id": "proceedings/wacv/2015/6683/0", "title": "2015 IEEE Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2020/01/08478369", "title": "Semi-Calibrated Photometric Stereo", "doi": null, "abstractUrl": "/journal/tp/2020/01/08478369/141AnpAbeCh", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500a317", "title": "Symmetric-light Photometric Stereo", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500a317/1B12PputGEg", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2019/3131/0/313100a574", "title": "Mobile Photometric Stereo with Keypoint-Based SLAM for Dense 3D Reconstruction", "doi": null, "abstractUrl": "/proceedings-article/3dv/2019/313100a574/1ezREwjZfFe", "parentPublication": { "id": "proceedings/3dv/2019/3131/0", "title": "2019 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2021/06/08943966", "title": "Single Day Outdoor Photometric Stereo", "doi": null, "abstractUrl": "/journal/tp/2021/06/08943966/1g6v5tld7IQ", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBBhN9J", "title": "Proceedings of the Workshop on Physics-Based Modeling in Computer Vision", "acronym": "pbmcv", "groupId": "1002324", "volume": "0", "displayVolume": "0", "year": "1995", "__typename": "ProceedingType" }, "article": { "id": "12OmNxbW4O4", "doi": "10.1109/PBMCV.1995.514684", "title": "Reflectance analysis under solar illumination", "normalizedTitle": "Reflectance analysis under solar illumination", "abstract": "The paper describes reflectance analysis of real images taken in an outdoor environment. In the past, the fundamental difficulties involved in analyzing outdoor images have prevented the use of traditional techniques for reflectance analysis such as shape from shading and photometric stereo. Those difficulties include: 1. multiple reflection components due to two light sources of different spectral and spatial distributions, namely the sunlight and the skylight, 2. ambiguity, in surface normal determination caused by the sun's restricted motion on the ecliptic, and 3. non-uniform surface albedo of object surfaces. In order to overcome those difficulties, we developed a reflectance model under solar illumination. Based on this model, we proposed solutions for each of these problems. We have successfully tested our algorithms using real color images taken both in a laboratory setup simulating sunlight, and in an outdoor environment under solar illumination.", "abstracts": [ { "abstractType": "Regular", "content": "The paper describes reflectance analysis of real images taken in an outdoor environment. In the past, the fundamental difficulties involved in analyzing outdoor images have prevented the use of traditional techniques for reflectance analysis such as shape from shading and photometric stereo. Those difficulties include: 1. multiple reflection components due to two light sources of different spectral and spatial distributions, namely the sunlight and the skylight, 2. ambiguity, in surface normal determination caused by the sun's restricted motion on the ecliptic, and 3. non-uniform surface albedo of object surfaces. In order to overcome those difficulties, we developed a reflectance model under solar illumination. Based on this model, we proposed solutions for each of these problems. We have successfully tested our algorithms using real color images taken both in a laboratory setup simulating sunlight, and in an outdoor environment under solar illumination.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The paper describes reflectance analysis of real images taken in an outdoor environment. In the past, the fundamental difficulties involved in analyzing outdoor images have prevented the use of traditional techniques for reflectance analysis such as shape from shading and photometric stereo. Those difficulties include: 1. multiple reflection components due to two light sources of different spectral and spatial distributions, namely the sunlight and the skylight, 2. ambiguity, in surface normal determination caused by the sun's restricted motion on the ecliptic, and 3. non-uniform surface albedo of object surfaces. In order to overcome those difficulties, we developed a reflectance model under solar illumination. Based on this model, we proposed solutions for each of these problems. We have successfully tested our algorithms using real color images taken both in a laboratory setup simulating sunlight, and in an outdoor environment under solar illumination.", "fno": "00514684", "keywords": [ "Image Processing", "Computer Vision", "Reflectance Analysis", "Solar Illumination", "Real Images", "Outdoor Environment", "Outdoor Images", "Shape From Shading", "Photometric Stereo", "Spatial Distributions", "Sunlight", "Skylight", "Real Color Images", "Reflectivity", "Lighting", "Image Analysis", "Shape", "Photometry", "Optical Reflection", "Light Sources", "Sun", "Testing", "Color" ], "authors": [ { "affiliation": "Robotics Inst., Carnegie Mellon Univ., Pittsburgh, PA, USA", "fullName": "Y. Sato", "givenName": "Y.", "surname": "Sato", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "K. Ikeuchi", "givenName": "K.", "surname": "Ikeuchi", "__typename": "ArticleAuthorType" } ], "idPrefix": "pbmcv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "1995-01-01T00:00:00", "pubType": "proceedings", "pages": "180,181,182,183,184,185,186,187", "year": "1995", "issn": null, "isbn": null, "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "00514683", "articleId": "12OmNAYGlsC", "__typename": "AdjacentArticleType" }, "next": null, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2009/3992/0/05206764", "title": "Illumination and spatially varying specular reflectance from a single view", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2009/05206764/12OmNARRYl8", "parentPublication": { "id": "proceedings/cvpr/2009/3992/0", "title": "2009 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/robot/1989/1938/0/00099963", "title": "Shape and reflectance from an image sequence generated using extended sources", "doi": null, "abstractUrl": "/proceedings-article/robot/1989/00099963/12OmNAq3hRr", "parentPublication": { "id": "proceedings/robot/1989/1938/0", "title": "1989 IEEE International Conference on Robotics and Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2009/4442/0/05457519", "title": "A common framework for ambient illumination in the dichromatic reflectance model", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2009/05457519/12OmNBDgZ0f", "parentPublication": { "id": "proceedings/iccvw/2009/4442/0", "title": "2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/1990/2062/1/00118069", "title": "Reconstructing shape from shading images under point light source illumination", "doi": null, "abstractUrl": "/proceedings-article/icpr/1990/00118069/12OmNBp52vd", "parentPublication": { "id": "proceedings/icpr/1990/2062/1", "title": "Proceedings 10th International Conference on Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/1989/1952/0/00037826", "title": "A theory of photometric stereo for a general class of reflectance maps", "doi": null, "abstractUrl": "/proceedings-article/cvpr/1989/00037826/12OmNqJHFLT", "parentPublication": { "id": "proceedings/cvpr/1989/1952/0", "title": "1989 IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/1990/2057/0/00139546", "title": "Simultaneous estimation of shape and reflectance maps from photometric stereo", "doi": null, "abstractUrl": "/proceedings-article/iccv/1990/00139546/12OmNrYlmHl", "parentPublication": { "id": "proceedings/iccv/1990/2057/0", "title": "Proceedings Third International Conference on Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2009/4420/0/05459381", "title": "Attached shadow coding: Estimating surface normals from shadows under unknown reflectance and lighting conditions", "doi": null, "abstractUrl": "/proceedings-article/iccv/2009/05459381/12OmNylsZWO", "parentPublication": { "id": "proceedings/iccv/2009/4420/0", "title": "2009 IEEE 12th International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2008/2174/0/04761892", "title": "Analysis of subsurface scattering under generic illumination", "doi": null, "abstractUrl": "/proceedings-article/icpr/2008/04761892/12OmNzd7bV9", "parentPublication": { "id": "proceedings/icpr/2008/2174/0", "title": "ICPR 2008 19th International Conference on Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2022/05/09712406", "title": "Predicting Surface Reflectance Properties of Outdoor Scenes Under Unknown Natural Illumination", "doi": null, "abstractUrl": "/magazine/cg/2022/05/09712406/1AZLEpMIeME", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2023/02/09557800", "title": "Neural Reflectance Capture in the View-Illumination Domain", "doi": null, "abstractUrl": "/journal/tg/2023/02/09557800/1xquQN6emfS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyKJiwN", "title": "2013 IEEE International Conference on Computer Vision Workshops (ICCVW)", "acronym": "iccvw", "groupId": "1800041", "volume": "0", "displayVolume": "0", "year": "2013", "__typename": "ProceedingType" }, "article": { "id": "12OmNyo1nYo", "doi": "10.1109/ICCVW.2013.124", "title": "Verification of Sky Models for Image Calibration", "normalizedTitle": "Verification of Sky Models for Image Calibration", "abstract": "Perception systems operating in outdoor scenarios face challenges due to the high dynamic range of the image, as different regions are illuminated by varying amounts of sunlight and skylight. A pre-processing step like image calibration can be used to convert the pixel values to an illumination independent domain such as reflectance. Each pixel is therefore represented by a characteristic material description, instead of an illumination and viewpoint dependent pixel colour. This assists object identification, segmentation and classification algorithms. This paper investigates modelling the sky colour through a number of parametric approaches typically used in the computer graphics community for rendering purposes, namely those developed by Preetham and Hosek-Wilkie. The models are compared in terms of chromaticity with observations taken from a camera and are used to develop an environment map for the application of inverse reflectometry of diffuse objects in an outdoor environment. This is of particular importance for applications involving imaging objects whose primary illumination source is skylight. It was found that the Hosek-Wilkie model produced more robust estimations and was less sensitive to changes in azimuth, while both models had similar reconstruction results with angular errors of approximately 0.15 radians.", "abstracts": [ { "abstractType": "Regular", "content": "Perception systems operating in outdoor scenarios face challenges due to the high dynamic range of the image, as different regions are illuminated by varying amounts of sunlight and skylight. A pre-processing step like image calibration can be used to convert the pixel values to an illumination independent domain such as reflectance. Each pixel is therefore represented by a characteristic material description, instead of an illumination and viewpoint dependent pixel colour. This assists object identification, segmentation and classification algorithms. This paper investigates modelling the sky colour through a number of parametric approaches typically used in the computer graphics community for rendering purposes, namely those developed by Preetham and Hosek-Wilkie. The models are compared in terms of chromaticity with observations taken from a camera and are used to develop an environment map for the application of inverse reflectometry of diffuse objects in an outdoor environment. This is of particular importance for applications involving imaging objects whose primary illumination source is skylight. It was found that the Hosek-Wilkie model produced more robust estimations and was less sensitive to changes in azimuth, while both models had similar reconstruction results with angular errors of approximately 0.15 radians.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Perception systems operating in outdoor scenarios face challenges due to the high dynamic range of the image, as different regions are illuminated by varying amounts of sunlight and skylight. A pre-processing step like image calibration can be used to convert the pixel values to an illumination independent domain such as reflectance. Each pixel is therefore represented by a characteristic material description, instead of an illumination and viewpoint dependent pixel colour. This assists object identification, segmentation and classification algorithms. This paper investigates modelling the sky colour through a number of parametric approaches typically used in the computer graphics community for rendering purposes, namely those developed by Preetham and Hosek-Wilkie. The models are compared in terms of chromaticity with observations taken from a camera and are used to develop an environment map for the application of inverse reflectometry of diffuse objects in an outdoor environment. This is of particular importance for applications involving imaging objects whose primary illumination source is skylight. It was found that the Hosek-Wilkie model produced more robust estimations and was less sensitive to changes in azimuth, while both models had similar reconstruction results with angular errors of approximately 0.15 radians.", "fno": "3022a907", "keywords": [ "Image Color Analysis", "Lighting", "Mathematical Model", "Computational Modeling", "Cameras", "Sun", "Sensors" ], "authors": [ { "affiliation": null, "fullName": "Rishi Ramakrishnan", "givenName": "Rishi", "surname": "Ramakrishnan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Juan Nieto", "givenName": "Juan", "surname": "Nieto", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Steve Scheding", "givenName": "Steve", "surname": "Scheding", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccvw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2013-12-01T00:00:00", "pubType": "proceedings", "pages": "907-914", "year": "2013", "issn": null, "isbn": "978-1-4799-3022-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3022a899", "articleId": "12OmNyRxFCh", "__typename": "AdjacentArticleType" }, "next": { "fno": "3022a915", "articleId": "12OmNBSSVme", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/imvip/2011/0230/0/06167859", "title": "Real-Time Diffuse Behavior Detection of Pixels from Outdoor Image Sequence", "doi": null, "abstractUrl": "/proceedings-article/imvip/2011/06167859/12OmNCcbE4U", "parentPublication": { "id": "proceedings/imvip/2011/0230/0", "title": "2011 Irish Machine Vision and Image Processing Conference (IMVIP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2012/2216/0/06460936", "title": "Probabilistic invariant image representation and associated distance measure", "doi": null, "abstractUrl": "/proceedings-article/icpr/2012/06460936/12OmNs4S8K9", "parentPublication": { "id": "proceedings/icpr/2012/2216/0", "title": "2012 21st International Conference on Pattern Recognition (ICPR 2012)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pbmcv/1995/7021/0/00514684", "title": "Reflectance analysis under solar illumination", "doi": null, "abstractUrl": "/proceedings-article/pbmcv/1995/00514684/12OmNxbW4O4", "parentPublication": { "id": "proceedings/pbmcv/1995/7021/0", "title": "Proceedings of the Workshop on Physics-Based Modeling in Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2017/0457/0/0457c373", "title": "Deep outdoor illumination estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457c373/12OmNy1SFMf", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2010/9343/0/05643558", "title": "Foreground and shadow occlusion handling for outdoor augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2010/05643558/12OmNyRPgDK", "parentPublication": { "id": "proceedings/ismar/2010/9343/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2013/3022/0/3022a899", "title": "Colour Constancy from Both Sides of the Shadow Edge", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2013/3022a899/12OmNyRxFCh", "parentPublication": { "id": "proceedings/iccvw/2013/3022/0", "title": "2013 IEEE International Conference on Computer Vision Workshops (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ams/2010/4062/0/4062a470", "title": "Interaction between Sunlight and the Sky Colour with 3D Objects in the Outdoor Virtual Environment", "doi": null, "abstractUrl": "/proceedings-article/ams/2010/4062a470/12OmNzC5SXJ", "parentPublication": { "id": "proceedings/ams/2010/4062/0", "title": "Asia International Conference on Modelling &amp; Simulation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2015/8332/0/8332a460", "title": "Shadow Detection and Sun Direction in Photo Collections", "doi": null, "abstractUrl": "/proceedings-article/3dv/2015/8332a460/12OmNzwpUq0", "parentPublication": { "id": "proceedings/3dv/2015/8332/0", "title": "2015 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/03/07867820", "title": "A Context-Aware Method for Authentically Simulating Outdoors Shadows for Mobile Augmented Reality", "doi": null, "abstractUrl": "/journal/tg/2018/03/07867820/13rRUy2YLT6", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040573", "title": "Online Tracking of Outdoor Lighting Variations for Augmented Reality with Moving Cameras", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040573/13rRUyY28Yr", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBkfRhw", "title": "2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNyuya6K", "doi": "10.1109/CVPR.2015.7299082", "title": "One-day outdoor photometric stereo via skylight estimation", "normalizedTitle": "One-day outdoor photometric stereo via skylight estimation", "abstract": "We present an outdoor photometric stereo method using images captured in a single day. We simulate a sky hemisphere for each image according to its GPS and timestamp, and parameterize the obtained sky hemisphere into a quadratic skylight and a Gaussian sunlight distribution. Unlike previous works which usually model outdoor illumination as a sum of constant ambient light and a distant point light, our method models natural illumination according to a popular sky model and thus provides sufficient constraints for shape reconstruction from one day images. We generate pixel profiles of uniformly sampled unit vectors for the corresponding time of captures and evaluate them using correlation with the actual pixel profiles. The estimated surface normal is refined by MRF optimization. We have tested our method to recover objects and scenes of various sizes in real-world outdoor daylight.", "abstracts": [ { "abstractType": "Regular", "content": "We present an outdoor photometric stereo method using images captured in a single day. We simulate a sky hemisphere for each image according to its GPS and timestamp, and parameterize the obtained sky hemisphere into a quadratic skylight and a Gaussian sunlight distribution. Unlike previous works which usually model outdoor illumination as a sum of constant ambient light and a distant point light, our method models natural illumination according to a popular sky model and thus provides sufficient constraints for shape reconstruction from one day images. We generate pixel profiles of uniformly sampled unit vectors for the corresponding time of captures and evaluate them using correlation with the actual pixel profiles. The estimated surface normal is refined by MRF optimization. We have tested our method to recover objects and scenes of various sizes in real-world outdoor daylight.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We present an outdoor photometric stereo method using images captured in a single day. We simulate a sky hemisphere for each image according to its GPS and timestamp, and parameterize the obtained sky hemisphere into a quadratic skylight and a Gaussian sunlight distribution. Unlike previous works which usually model outdoor illumination as a sum of constant ambient light and a distant point light, our method models natural illumination according to a popular sky model and thus provides sufficient constraints for shape reconstruction from one day images. We generate pixel profiles of uniformly sampled unit vectors for the corresponding time of captures and evaluate them using correlation with the actual pixel profiles. The estimated surface normal is refined by MRF optimization. We have tested our method to recover objects and scenes of various sizes in real-world outdoor daylight.", "fno": "07299082", "keywords": [], "authors": [ { "affiliation": "Robotics and Computer Vision Lab, KAIST, Yuseong-gu, Daejeon, South Korea", "fullName": "Jiyoung Jung", "givenName": "Jiyoung", "surname": "Jung", "__typename": "ArticleAuthorType" }, { "affiliation": "Robotics and Computer Vision Lab, KAIST, Yuseong-gu, Daejeon, South Korea", "fullName": "Joon-Young Lee", "givenName": "Joon-Young", "surname": "Lee", "__typename": "ArticleAuthorType" }, { "affiliation": "Robotics and Computer Vision Lab, KAIST, Yuseong-gu, Daejeon, South Korea", "fullName": "In So Kweon", "givenName": "In So", "surname": "Kweon", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-06-01T00:00:00", "pubType": "proceedings", "pages": "4521-4529", "year": "2015", "issn": "1063-6919", "isbn": "978-1-4673-6964-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07299081", "articleId": "12OmNAWH9Fl", "__typename": "AdjacentArticleType" }, "next": { "fno": "07299083", "articleId": "12OmNCwCLsN", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccp/2015/8667/0/07168379", "title": "What Is a Good Day for Outdoor Photometric Stereo?", "doi": null, "abstractUrl": "/proceedings-article/iccp/2015/07168379/12OmNBmf3by", "parentPublication": { "id": "proceedings/iccp/2015/8667/0", "title": "2015 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2015/8332/0/8332a028", "title": "x-Hour Outdoor Photometric Stereo", "doi": null, "abstractUrl": "/proceedings-article/3dv/2015/8332a028/12OmNCesr5K", "parentPublication": { "id": "proceedings/3dv/2015/8332/0", "title": "2015 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2009/4420/0/05459163", "title": "Estimating natural illumination from a single outdoor image", "doi": null, "abstractUrl": "/proceedings-article/iccv/2009/05459163/12OmNrJiD1c", "parentPublication": { "id": "proceedings/iccv/2009/4420/0", "title": "2009 IEEE 12th International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2013/6463/0/06528306", "title": "Outdoor photometric stereo", "doi": null, "abstractUrl": "/proceedings-article/iccp/2013/06528306/12OmNrNh0Dh", "parentPublication": { "id": "proceedings/iccp/2013/6463/0", "title": "2013 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2012/1226/0/034P1A34", "title": "Photometric stereo for outdoor webcams", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2012/034P1A34/12OmNrY3LEd", "parentPublication": { "id": "proceedings/cvpr/2012/1226/0", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200p5293", "title": "Hierarchical Disentangled Representation Learning for Outdoor Illumination Estimation and Editing", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200p5293/1BmEEvuyjUQ", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2022/5670/0/567000a690", "title": "Polar-Photometric Stereo Under Natural Illumination", "doi": null, "abstractUrl": "/proceedings-article/3dv/2022/567000a690/1KYsvRkYv04", "parentPublication": { "id": "proceedings/3dv/2022/5670/0", "title": "2022 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2021/06/08943966", "title": "Single Day Outdoor Photometric Stereo", "doi": null, "abstractUrl": "/journal/tp/2021/06/08943966/1g6v5tld7IQ", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300g920", "title": "Deep Sky Modeling for Single Image Outdoor Lighting Estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300g920/1gyrdbEY2sM", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300k0150", "title": "All-Weather Deep Outdoor Lighting Estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300k0150/1gyrg6Ricuc", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "17D45VtKiso", "title": "2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "acronym": "sibgrapi", "groupId": "1000131", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "17D45XvMceO", "doi": "10.1109/SIBGRAPI.2018.00048", "title": "Asynchronous Stroboscopic Structured Lighting Image Processing Using Low-Cost Cameras", "normalizedTitle": "Asynchronous Stroboscopic Structured Lighting Image Processing Using Low-Cost Cameras", "abstract": "Structured lighting (SL) image processing relies on the generation of known illumination patterns synchronized with the camera frame rate and is commonly implemented using syncing capable cameras. In general, such cameras employ global shutters, that exposes the whole frame at once. However, most modern digital cameras use rolling shutters, which expose each line at different intervals, impairing most structured lighting applications. In this paper we introduce an asynchronous SL technique that can be used by any rolling shutter digital camera. While the use of stroboscopic illumination partially solves for the line exposure shift, the phase difference between the camera and lighting clocks results in stripe artifacts that move vertically in the video stream. These stripes are detected and tracked using a Kalman filter. Two asynchronous stroboscopic SL methods are proposed. The first method, image differencing, minimizes the stripe artifacts. The second method, image compositing, completely removes the artifacts. We demonstrate the use of the asynchronous differential lighting technique in a pupil detector using a low-cost high-speed camera with no synchronization means, with the lighting running independently at a higher, unknown frequency to the application.", "abstracts": [ { "abstractType": "Regular", "content": "Structured lighting (SL) image processing relies on the generation of known illumination patterns synchronized with the camera frame rate and is commonly implemented using syncing capable cameras. In general, such cameras employ global shutters, that exposes the whole frame at once. However, most modern digital cameras use rolling shutters, which expose each line at different intervals, impairing most structured lighting applications. In this paper we introduce an asynchronous SL technique that can be used by any rolling shutter digital camera. While the use of stroboscopic illumination partially solves for the line exposure shift, the phase difference between the camera and lighting clocks results in stripe artifacts that move vertically in the video stream. These stripes are detected and tracked using a Kalman filter. Two asynchronous stroboscopic SL methods are proposed. The first method, image differencing, minimizes the stripe artifacts. The second method, image compositing, completely removes the artifacts. We demonstrate the use of the asynchronous differential lighting technique in a pupil detector using a low-cost high-speed camera with no synchronization means, with the lighting running independently at a higher, unknown frequency to the application.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Structured lighting (SL) image processing relies on the generation of known illumination patterns synchronized with the camera frame rate and is commonly implemented using syncing capable cameras. In general, such cameras employ global shutters, that exposes the whole frame at once. However, most modern digital cameras use rolling shutters, which expose each line at different intervals, impairing most structured lighting applications. In this paper we introduce an asynchronous SL technique that can be used by any rolling shutter digital camera. While the use of stroboscopic illumination partially solves for the line exposure shift, the phase difference between the camera and lighting clocks results in stripe artifacts that move vertically in the video stream. These stripes are detected and tracked using a Kalman filter. Two asynchronous stroboscopic SL methods are proposed. The first method, image differencing, minimizes the stripe artifacts. The second method, image compositing, completely removes the artifacts. We demonstrate the use of the asynchronous differential lighting technique in a pupil detector using a low-cost high-speed camera with no synchronization means, with the lighting running independently at a higher, unknown frequency to the application.", "fno": "926400a321", "keywords": [ "Cameras", "Image Processing", "Kalman Filters", "Lighting", "Object Detection", "Object Tracking", "Stroboscopes", "Video Signal Processing", "Asynchronous Stroboscopic Structured Lighting Image Processing", "Camera Frame Rate", "Structured Lighting Applications", "Video Stream", "Lighting Clocks", "Digital Cameras", "High Speed Camera", "Asynchronous Differential Lighting Technique", "Image Compositing", "Image Differencing", "Asynchronous Stroboscopic SL Methods", "Stripe Artifacts", "Phase Difference", "Line Exposure Shift", "Stroboscopic Illumination", "Rolling Shutter Digital Camera", "Asynchronous SL Technique", "Lighting", "Synchronization", "Frequency Synchronization", "Clocks", "Software", "Digital Cameras", "Structured Lighting", "Asynchronous Stroboscopic Structured Lighting", "Rolling Shutter", "Pupil Detection" ], "authors": [ { "affiliation": null, "fullName": "Frank Helbert Borsato", "givenName": "Frank Helbert", "surname": "Borsato", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Carlos Hitoshi Morimoto", "givenName": "Carlos Hitoshi", "surname": "Morimoto", "__typename": "ArticleAuthorType" } ], "idPrefix": "sibgrapi", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-10-01T00:00:00", "pubType": "proceedings", "pages": "321-328", "year": "2018", "issn": null, "isbn": "978-1-5386-9264-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "926400a313", "articleId": "17D45VVho41", "__typename": "AdjacentArticleType" }, "next": { "fno": "926400a329", "articleId": "17D45XwUAJr", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ssiai/2014/4053/0/06806031", "title": "Low-cost camera array for mitigating lighting range effects", "doi": null, "abstractUrl": "/proceedings-article/ssiai/2014/06806031/12OmNAT0mLN", "parentPublication": { "id": "proceedings/ssiai/2014/4053/0", "title": "2014 IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2017/2219/0/2219a015", "title": "Building Structured Lighting Applications Using Low-Cost Cameras", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2017/2219a015/12OmNBIFmto", "parentPublication": { "id": "proceedings/sibgrapi/2017/2219/0", "title": "2017 30th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2012/1611/0/06239343", "title": "Recovering spectral reflectance under commonly available lighting conditions", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2012/06239343/12OmNBTs7yq", "parentPublication": { "id": "proceedings/cvprw/2012/1611/0", "title": "2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/crv/2011/4362/0/05957553", "title": "Adaptive Lighting for Machine Vision Applications", "doi": null, "abstractUrl": "/proceedings-article/crv/2011/05957553/12OmNC1Gufk", "parentPublication": { "id": "proceedings/crv/2011/4362/0", "title": "2011 Canadian Conference on Computer and Robot Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2013/3022/0/3022a844", "title": "Quick Approximation of Camera's Spectral Response from Casual Lighting", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2013/3022a844/12OmNvDZEZn", "parentPublication": { "id": "proceedings/iccvw/2013/3022/0", "title": "2013 IEEE International Conference on Computer Vision Workshops (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2015/9711/0/5720a502", "title": "A Fast and Accurate Eye Tracker Using Stroboscopic Differential Lighting", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2015/5720a502/12OmNzC5Tmm", "parentPublication": { "id": "proceedings/iccvw/2015/9711/0", "title": "2015 IEEE International Conference on Computer Vision Workshop (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/04/v0652", "title": "Tabletop Computed Lighting for Practical Digital Photography", "doi": null, "abstractUrl": "/journal/tg/2007/04/v0652/13rRUILtJzp", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sitis/2022/6495/0/649500a281", "title": "Lighting Spectral Power Distribution Estimation With RGB Camera", "doi": null, "abstractUrl": "/proceedings-article/sitis/2022/649500a281/1MeoGuSgRPO", "parentPublication": { "id": "proceedings/sitis/2022/6495/0", "title": "2022 16th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csci/2018/1360/0/136000b448", "title": "A New Proposal of Smart Lighting System Based on Radar and Camera Sensors for Smart City", "doi": null, "abstractUrl": "/proceedings-article/csci/2018/136000b448/1gjRzmLfXxu", "parentPublication": { "id": "proceedings/csci/2018/1360/0", "title": "2018 International Conference on Computational Science and Computational Intelligence (CSCI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900o4755", "title": "Indoor Lighting Estimation using an Event Camera", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900o4755/1yeMhx0N8OI", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzsJ7D3", "title": "37th Annual IEEE Conference on Local Computer Networks", "acronym": "lcn", "groupId": "1000419", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNqzcvC0", "doi": "10.1109/LCN.2012.6423586", "title": "3DSVAT: A 3D Stereoscopic Vulnerability Assessment Tool for network security", "normalizedTitle": "3DSVAT: A 3D Stereoscopic Vulnerability Assessment Tool for network security", "abstract": "As the volume of network data continues to increase and networks become more complex, the ability to accurately manage and analyze data quickly becomes a difficult problem. Many network management tools already use two-dimensional (2D) and three-dimensional (3D) visualization techniques to help support decision-making and reasoning of network anomalies and activity. However, a poor user interface combined with the massive amount of data could obfuscate important network details. As a result, administrators may fail to detect and identify malicious network behavior in a timely manner. 3D visualizations address this challenge by introducing monocular and binocular visual cues to portray depth and to increase the perceived viewing area. In this work, we explore these cues for 3D network security applications, with a particular emphasis on binocular disparity or stereoscopic 3D. Currently, no network security tool takes advantage of the enhanced depth perception provided by stereoscopic 3D technologies for vulnerability assessment. Compared to traditional 3D systems, stereoscopic 3D helps improve the perception of depth, which can, in turn reduce the number of errors and increase response times of network administrators. Thus, we introduce a stereoscopic 3D visual Framework for Rendering Enhanced 3D Stereoscopic Visualizations for Network Security (FRE3DS). Our novel framework uses state-of-the art 3D graphics rendering to assist in 3D visualizations for network security applications. Moreover, utilizing our framework, we propose a new 3D Stereoscopic Vulnerability Assessment Tool (3DSVAT). We illustrate the use of 3DSVAT to assist in rapid detection and correlation of attack vulnerabilities in a subset of a modified local area network data set using the enhanced perception of depth in a stereoscopic 3D environment.", "abstracts": [ { "abstractType": "Regular", "content": "As the volume of network data continues to increase and networks become more complex, the ability to accurately manage and analyze data quickly becomes a difficult problem. Many network management tools already use two-dimensional (2D) and three-dimensional (3D) visualization techniques to help support decision-making and reasoning of network anomalies and activity. However, a poor user interface combined with the massive amount of data could obfuscate important network details. As a result, administrators may fail to detect and identify malicious network behavior in a timely manner. 3D visualizations address this challenge by introducing monocular and binocular visual cues to portray depth and to increase the perceived viewing area. In this work, we explore these cues for 3D network security applications, with a particular emphasis on binocular disparity or stereoscopic 3D. Currently, no network security tool takes advantage of the enhanced depth perception provided by stereoscopic 3D technologies for vulnerability assessment. Compared to traditional 3D systems, stereoscopic 3D helps improve the perception of depth, which can, in turn reduce the number of errors and increase response times of network administrators. Thus, we introduce a stereoscopic 3D visual Framework for Rendering Enhanced 3D Stereoscopic Visualizations for Network Security (FRE3DS). Our novel framework uses state-of-the art 3D graphics rendering to assist in 3D visualizations for network security applications. Moreover, utilizing our framework, we propose a new 3D Stereoscopic Vulnerability Assessment Tool (3DSVAT). We illustrate the use of 3DSVAT to assist in rapid detection and correlation of attack vulnerabilities in a subset of a modified local area network data set using the enhanced perception of depth in a stereoscopic 3D environment.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "As the volume of network data continues to increase and networks become more complex, the ability to accurately manage and analyze data quickly becomes a difficult problem. Many network management tools already use two-dimensional (2D) and three-dimensional (3D) visualization techniques to help support decision-making and reasoning of network anomalies and activity. However, a poor user interface combined with the massive amount of data could obfuscate important network details. As a result, administrators may fail to detect and identify malicious network behavior in a timely manner. 3D visualizations address this challenge by introducing monocular and binocular visual cues to portray depth and to increase the perceived viewing area. In this work, we explore these cues for 3D network security applications, with a particular emphasis on binocular disparity or stereoscopic 3D. Currently, no network security tool takes advantage of the enhanced depth perception provided by stereoscopic 3D technologies for vulnerability assessment. Compared to traditional 3D systems, stereoscopic 3D helps improve the perception of depth, which can, in turn reduce the number of errors and increase response times of network administrators. Thus, we introduce a stereoscopic 3D visual Framework for Rendering Enhanced 3D Stereoscopic Visualizations for Network Security (FRE3DS). Our novel framework uses state-of-the art 3D graphics rendering to assist in 3D visualizations for network security applications. Moreover, utilizing our framework, we propose a new 3D Stereoscopic Vulnerability Assessment Tool (3DSVAT). We illustrate the use of 3DSVAT to assist in rapid detection and correlation of attack vulnerabilities in a subset of a modified local area network data set using the enhanced perception of depth in a stereoscopic 3D environment.", "fno": "06423586", "keywords": [ "Stereo Image Processing", "Data Visualization", "Security", "Visualization", "Cameras", "Rendering Computer Graphics", "Arrays", "Vulnerability Assessment Tool", "Stereoscopic 3 D", "Security Visualization" ], "authors": [ { "affiliation": "CSC Lab., The School of ECE, Georgia Institute of Technology, Atlanta, 30332, USA", "fullName": "Troy Nunnally", "givenName": "Troy", "surname": "Nunnally", "__typename": "ArticleAuthorType" }, { "affiliation": "GT CAP Group, The School of ECE, Georgia Institute of Technology, Atlanta, 30332, USA", "fullName": "A. Selcuk Uluagac", "givenName": "A. Selcuk", "surname": "Uluagac", "__typename": "ArticleAuthorType" }, { "affiliation": "CSC Lab., The School of ECE, Georgia Institute of Technology, Atlanta, 30332, USA", "fullName": "John A. Copeland", "givenName": "John A.", "surname": "Copeland", "__typename": "ArticleAuthorType" }, { "affiliation": "GT CAP Group, The School of ECE, Georgia Institute of Technology, Atlanta, 30332, USA", "fullName": "Raheem Beyah", "givenName": "Raheem", "surname": "Beyah", "__typename": "ArticleAuthorType" } ], "idPrefix": "lcn", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-10-01T00:00:00", "pubType": "proceedings", "pages": "111-118", "year": "2012", "issn": "0742-1303", "isbn": "978-1-4673-1565-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06423585", "articleId": "12OmNCwlafJ", "__typename": "AdjacentArticleType" }, "next": { "fno": "06423587", "articleId": "12OmNx0A7Po", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2012/1226/0/007P1A07", "title": "Scene warping: Layer-based stereoscopic image resizing", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2012/007P1A07/12OmNAiFI8D", "parentPublication": { "id": "proceedings/cvpr/2012/1226/0", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgames/2011/1451/0/06000331", "title": "Stereoscopic 3D in video games: A review of current design practices and challenges", "doi": null, "abstractUrl": "/proceedings-article/cgames/2011/06000331/12OmNqJZgHD", "parentPublication": { "id": "proceedings/cgames/2011/1451/0", "title": "2011 16th International Conference on Computer Games (CGAMES)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2016/7258/0/07552873", "title": "Learning-based quality assessment of retargeted stereoscopic images", "doi": null, "abstractUrl": "/proceedings-article/icme/2016/07552873/12OmNrFkeQd", "parentPublication": { "id": "proceedings/icme/2016/7258/0", "title": "2016 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dvis/2014/6826/0/07160093", "title": "The (possible) utility of stereoscopic 3D displays for information visualization: The good, the bad, and the ugly", "doi": null, "abstractUrl": "/proceedings-article/3dvis/2014/07160093/12OmNxWcHag", "parentPublication": { "id": "proceedings/3dvis/2014/6826/0", "title": "2014 IEEE VIS International Workshop on 3DVis (3DVis)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iih-msp/2011/4517/0/4517a081", "title": "Perceptual Watermarking for 3D Stereoscopic Video Using Depth Information", "doi": null, "abstractUrl": "/proceedings-article/iih-msp/2011/4517a081/12OmNxYbT0j", "parentPublication": { "id": "proceedings/iih-msp/2011/4517/0", "title": "Intelligent Information Hiding and Multimedia Signal Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2017/0560/0/08026266", "title": "Quality assessment of stereoscopic 3D images based on local and global visual characteristics", "doi": null, "abstractUrl": "/proceedings-article/icmew/2017/08026266/12OmNzC5TqJ", "parentPublication": { "id": "proceedings/icmew/2017/0560/0", "title": "2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/01/06824802", "title": "Geometrically Consistent Stereoscopic Image Editing Using Patch-Based Synthesis", "doi": null, "abstractUrl": "/journal/tg/2015/01/06824802/13rRUwInvsT", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2013/08/ttg2013081288", "title": "Changing Perspective in Stereoscopic Images", "doi": null, "abstractUrl": "/journal/tg/2013/08/ttg2013081288/13rRUwghd4Y", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2011/12/ttg2011122325", "title": "Stereoscopic Highlighting: 2D Graph Visualization on Stereo Displays", "doi": null, "abstractUrl": "/journal/tg/2011/12/ttg2011122325/13rRUy2YLSZ", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2019/1975/0/197500b655", "title": "Warping-Based Stereoscopic 3D Video Retargeting With Depth Remapping", "doi": null, "abstractUrl": "/proceedings-article/wacv/2019/197500b655/18j8LvV2AJG", "parentPublication": { "id": "proceedings/wacv/2019/1975/0", "title": "2019 IEEE Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyUWQR6", "title": "Virtual Reality Annual International Symposium", "acronym": "vrais", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "1993", "__typename": "ProceedingType" }, "article": { "id": "12OmNrJAdLH", "doi": "10.1109/VRAIS.1993.380802", "title": "The effect of interocular distance upon operator performance using stereoscopic displays to perform virtual depth tasks", "normalizedTitle": "The effect of interocular distance upon operator performance using stereoscopic displays to perform virtual depth tasks", "abstract": "When presenting virtual images to a user performing a simple task requiring depth perception, the use of stereoscopic projections results in a ten-fold reduction in mean alignment error as compared to the use of monocular projections. Although average physiological interocular distance is 6.3 cm, it is found that any interocular distance of greater than 3 cm used in the stereo projection model is adequate to provide a user with maximal performance in the depth perception task. No statistically significant increase in performance can be correlated to increasing interocular distances greater than 3 cm. Since it is often beneficial to reduce retinal disparity between the left and right images to increase the presentable depth range, reduce image fusion problems, and reduce operator fatigue, these results suggests that smaller than physiological interocular distances should be considered when implementing a stereoscopic vision system for virtual environments and telepresence systems.", "abstracts": [ { "abstractType": "Regular", "content": "When presenting virtual images to a user performing a simple task requiring depth perception, the use of stereoscopic projections results in a ten-fold reduction in mean alignment error as compared to the use of monocular projections. Although average physiological interocular distance is 6.3 cm, it is found that any interocular distance of greater than 3 cm used in the stereo projection model is adequate to provide a user with maximal performance in the depth perception task. No statistically significant increase in performance can be correlated to increasing interocular distances greater than 3 cm. Since it is often beneficial to reduce retinal disparity between the left and right images to increase the presentable depth range, reduce image fusion problems, and reduce operator fatigue, these results suggests that smaller than physiological interocular distances should be considered when implementing a stereoscopic vision system for virtual environments and telepresence systems.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "When presenting virtual images to a user performing a simple task requiring depth perception, the use of stereoscopic projections results in a ten-fold reduction in mean alignment error as compared to the use of monocular projections. Although average physiological interocular distance is 6.3 cm, it is found that any interocular distance of greater than 3 cm used in the stereo projection model is adequate to provide a user with maximal performance in the depth perception task. No statistically significant increase in performance can be correlated to increasing interocular distances greater than 3 cm. Since it is often beneficial to reduce retinal disparity between the left and right images to increase the presentable depth range, reduce image fusion problems, and reduce operator fatigue, these results suggests that smaller than physiological interocular distances should be considered when implementing a stereoscopic vision system for virtual environments and telepresence systems.", "fno": "00380802", "keywords": [ "Telepresence", "Interocular Distance", "Operator Performance", "Stereoscopic Displays", "Virtual Depth Tasks", "Virtual Images", "Depth Perception", "Mean Alignment Error", "Monocular Projections", "Physiological Interocular Distance", "Stereo Projection Model", "Retinal Disparity", "Image Fusion Problems", "Operator Fatigue", "Stereoscopic Vision System", "Virtual Environments" ], "authors": [ { "affiliation": "Center for Design Res., Stanford Univ., CA, USA", "fullName": "L.B. Rosenberg", "givenName": "L.B.", "surname": "Rosenberg", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrais", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1993-09-01T00:00:00", "pubType": "proceedings", "pages": "27-32", "year": "1993", "issn": null, "isbn": "0-7803-1363-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "00380801", "articleId": "12OmNylboFY", "__typename": "AdjacentArticleType" }, "next": { "fno": "00380803", "articleId": "12OmNCd2rG7", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNz5JC0v", "title": "2010 International Conference on Digital Image Computing: Techniques and Applications", "acronym": "dicta", "groupId": "1001512", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNwE9Ot5", "doi": "10.1109/DICTA.2010.49", "title": "A Two-Stage Correlation Method for Stereoscopic Depth Estimation", "normalizedTitle": "A Two-Stage Correlation Method for Stereoscopic Depth Estimation", "abstract": "The computation of stereoscopic depth is an important field of computer vision. Although a large variety of algorithms has been developed, the traditional correlation-based versions of these algorithms are prevalent. This is mainly due to easy implementation and handling but also to the linear computational complexity, as compared to more elaborated algorithms based on diffusion processes, graph-cut or bilateral filtering. In this paper, we introduce a new two-stage matching cost for the traditional approach: the summed normalized cross-correlation (SNCC). This new cost function performs a normalized cross-correlation in the first stage and aggregates the correlation values in a second stage. We show that this new measure can be implemented efficiently and that it leads to a substantial improvement of the performance of the traditional stereo approach because it is less sensitive to high contrast outliers.", "abstracts": [ { "abstractType": "Regular", "content": "The computation of stereoscopic depth is an important field of computer vision. Although a large variety of algorithms has been developed, the traditional correlation-based versions of these algorithms are prevalent. This is mainly due to easy implementation and handling but also to the linear computational complexity, as compared to more elaborated algorithms based on diffusion processes, graph-cut or bilateral filtering. In this paper, we introduce a new two-stage matching cost for the traditional approach: the summed normalized cross-correlation (SNCC). This new cost function performs a normalized cross-correlation in the first stage and aggregates the correlation values in a second stage. We show that this new measure can be implemented efficiently and that it leads to a substantial improvement of the performance of the traditional stereo approach because it is less sensitive to high contrast outliers.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The computation of stereoscopic depth is an important field of computer vision. Although a large variety of algorithms has been developed, the traditional correlation-based versions of these algorithms are prevalent. This is mainly due to easy implementation and handling but also to the linear computational complexity, as compared to more elaborated algorithms based on diffusion processes, graph-cut or bilateral filtering. In this paper, we introduce a new two-stage matching cost for the traditional approach: the summed normalized cross-correlation (SNCC). This new cost function performs a normalized cross-correlation in the first stage and aggregates the correlation values in a second stage. We show that this new measure can be implemented efficiently and that it leads to a substantial improvement of the performance of the traditional stereo approach because it is less sensitive to high contrast outliers.", "fno": "4271a227", "keywords": [ "Stereoscopic Depth", "Cost Function", "Reduced Fattening" ], "authors": [ { "affiliation": null, "fullName": "Nils Einecke", "givenName": "Nils", "surname": "Einecke", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Julian Eggert", "givenName": "Julian", "surname": "Eggert", "__typename": "ArticleAuthorType" } ], "idPrefix": "dicta", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-12-01T00:00:00", "pubType": "proceedings", "pages": "227-234", "year": "2010", "issn": null, "isbn": "978-0-7695-4271-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4271a220", "articleId": "12OmNAnuTqJ", "__typename": "AdjacentArticleType" }, "next": { "fno": "4271a235", "articleId": "12OmNAYoKiT", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icise/2009/3887/0/pid981884", "title": "FPGA Based Parallel Architectures for Normalized Cross-Correlation", "doi": null, "abstractUrl": "/proceedings-article/icise/2009/pid981884/12OmNCbkQCv", "parentPublication": { "id": "proceedings/icise/2009/3887/0", "title": "Information Science and Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mvhi/2010/4009/0/4009a003", "title": "Hypercomplex Correlation for Defect Detection", "doi": null, "abstractUrl": "/proceedings-article/mvhi/2010/4009a003/12OmNCd2rqD", "parentPublication": { "id": "proceedings/mvhi/2010/4009/0", "title": "Machine Vision and Human-machine Interface, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iih-msp/2009/3762/0/3762a009", "title": "A Fast Template Matching Method for Rotation Invariance Using Two-Stage Process", "doi": null, "abstractUrl": "/proceedings-article/iih-msp/2009/3762a009/12OmNvxsSRc", "parentPublication": { "id": "proceedings/iih-msp/2009/3762/0", "title": "Intelligent Information Hiding and Multimedia Signal Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icycs/2008/3398/0/3398a831", "title": "Precise Depth Perception in Projective Stereoscopic Display", "doi": null, "abstractUrl": "/proceedings-article/icycs/2008/3398a831/12OmNwtEEM5", "parentPublication": { "id": "proceedings/icycs/2008/3398/0", "title": "2008 9th International Conference for Young Computer Scientists", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/1999/0185/0/01850080", "title": "A Technique for Precise Depth Representation in Stereoscopic Display", "doi": null, "abstractUrl": "/proceedings-article/cgi/1999/01850080/12OmNx3q6XQ", "parentPublication": { "id": "proceedings/cgi/1999/0185/0", "title": "Computer Graphics International Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icig/2011/4541/0/4541a249", "title": "Inter-view Coding Correlation Based Error Concealment for Stereoscopic Video Transmission", "doi": null, "abstractUrl": "/proceedings-article/icig/2011/4541a249/12OmNx4Q6CV", "parentPublication": { "id": "proceedings/icig/2011/4541/0", "title": "Image and Graphics, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pdp/2008/3089/0/3089a407", "title": "Exploiting Data- and Thread-Level Parallelism for Image Correlation", "doi": null, "abstractUrl": "/proceedings-article/pdp/2008/3089a407/12OmNxcdFYZ", "parentPublication": { "id": "proceedings/pdp/2008/3089/0", "title": "2008 16th Euromicro Conference on Parallel, Distributed and Network-based Processing - PDP '08", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aipr/2009/5146/0/05466291", "title": "Spatial-spectral cross correlation for reliable multispectral image registration", "doi": null, "abstractUrl": "/proceedings-article/aipr/2009/05466291/12OmNyfvpSD", "parentPublication": { "id": "proceedings/aipr/2009/5146/0", "title": "2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/1984/03/04767532", "title": "A Two-Stage Cross Correlation Approach to Template Matching", "doi": null, "abstractUrl": "/journal/tp/1984/03/04767532/13rRUwdIOVA", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200c249", "title": "High Quality Disparity Remapping with Two-Stage Warping", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200c249/1BmGb6C6fn2", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwbcJ4L", "title": "2008 9th International Conference for Young Computer Scientists", "acronym": "icycs", "groupId": "1002545", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNwtEEM5", "doi": "10.1109/ICYCS.2008.409", "title": "Precise Depth Perception in Projective Stereoscopic Display", "normalizedTitle": "Precise Depth Perception in Projective Stereoscopic Display", "abstract": "There is deviation between perceived depth andtheoretical depth of virtual object in stereoscopicvirtual environment. This paper analyzes its possiblecauses with optical geometry. In addition, a correctingmethod performed by setting synthetic viewpointdynamically is proposed. Experimental results showthat this method can reduce the depth deviation to lessthan 10 millimeter, meeting the requirements forlocating and direct manipulation in virtualenvironment.", "abstracts": [ { "abstractType": "Regular", "content": "There is deviation between perceived depth andtheoretical depth of virtual object in stereoscopicvirtual environment. This paper analyzes its possiblecauses with optical geometry. In addition, a correctingmethod performed by setting synthetic viewpointdynamically is proposed. Experimental results showthat this method can reduce the depth deviation to lessthan 10 millimeter, meeting the requirements forlocating and direct manipulation in virtualenvironment.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "There is deviation between perceived depth andtheoretical depth of virtual object in stereoscopicvirtual environment. This paper analyzes its possiblecauses with optical geometry. In addition, a correctingmethod performed by setting synthetic viewpointdynamically is proposed. Experimental results showthat this method can reduce the depth deviation to lessthan 10 millimeter, meeting the requirements forlocating and direct manipulation in virtualenvironment.", "fno": "3398a831", "keywords": [ "Virtual Reality", "Depth Perception", "Stereoscopic Display", "Head Tracking Display" ], "authors": [ { "affiliation": null, "fullName": "Liping Lin", "givenName": "Liping", "surname": "Lin", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Pingdong Wu", "givenName": "Pingdong", "surname": "Wu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jie Huang", "givenName": "Jie", "surname": "Huang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jian Li", "givenName": "Jian", "surname": "Li", "__typename": "ArticleAuthorType" } ], "idPrefix": "icycs", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-11-01T00:00:00", "pubType": "proceedings", "pages": "831-836", "year": "2008", "issn": null, "isbn": "978-0-7695-3398-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3398a826", "articleId": "12OmNvpewc3", "__typename": "AdjacentArticleType" }, "next": { "fno": "3398a837", "articleId": "12OmNzBOhtE", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/avss/2011/0844/0/06027293", "title": "Stereoscopic viewing facilitates the perception of crowds", "doi": null, "abstractUrl": "/proceedings-article/avss/2011/06027293/12OmNBEpnya", "parentPublication": { "id": "proceedings/avss/2011/0844/0", "title": "2011 8th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1993/1363/0/00380802", "title": "The effect of interocular distance upon operator performance using stereoscopic displays to perform virtual depth tasks", "doi": null, "abstractUrl": "/proceedings-article/vrais/1993/00380802/12OmNrJAdLH", "parentPublication": { "id": "proceedings/vrais/1993/1363/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dicta/2010/4271/0/4271a227", "title": "A Two-Stage Correlation Method for Stereoscopic Depth Estimation", "doi": null, "abstractUrl": "/proceedings-article/dicta/2010/4271a227/12OmNwE9Ot5", "parentPublication": { "id": "proceedings/dicta/2010/4271/0", "title": "2010 International Conference on Digital Image Computing: Techniques and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/computationworld/2009/3862/0/3862a460", "title": "Depth Perception within Virtual Environments: A Comparative Study Between Wide Screen Stereoscopic Displays and Head Mounted Devices", "doi": null, "abstractUrl": "/proceedings-article/computationworld/2009/3862a460/12OmNwJPN1e", "parentPublication": { "id": "proceedings/computationworld/2009/3862/0", "title": "Future Computing, Service Computation, Cognitive, Adaptive, Content, Patterns, Computation World", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/1999/0185/0/01850080", "title": "A Technique for Precise Depth Representation in Stereoscopic Display", "doi": null, "abstractUrl": "/proceedings-article/cgi/1999/01850080/12OmNx3q6XQ", "parentPublication": { "id": "proceedings/cgi/1999/0185/0", "title": "Computer Graphics International Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/smc-it/2009/3637/0/3637a143", "title": "Motor Activity-Perception Based Approach for Improving Teleoperation Systems", "doi": null, "abstractUrl": "/proceedings-article/smc-it/2009/3637a143/12OmNxFaLz6", "parentPublication": { "id": "proceedings/smc-it/2009/3637/0", "title": "2009 Third IEEE International Conference on Space Mission Challenges for Information Technology (SMC-IT 2009)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icic/2009/3634/4/3634d399", "title": "Stereoscopic Display on Post-Process of ANSYS", "doi": null, "abstractUrl": "/proceedings-article/icic/2009/3634d399/12OmNxFsmAR", "parentPublication": { "id": "proceedings/icic/2009/3634/4", "title": "2009 Second International Conference on Information and Computing Science", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2002/1492/0/14920081", "title": "A Testbed for Precise Registration, Natural Occlusion and Interaction in an Augmented Environment Using a Head-Mounted Projective Display (HMPD)", "doi": null, "abstractUrl": "/proceedings-article/vr/2002/14920081/12OmNylboC4", "parentPublication": { "id": "proceedings/vr/2002/1492/0", "title": "Proceedings IEEE Virtual Reality 2002", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2002/02/v0129", "title": "A Geometric Comparison of Algorithms for Fusion Control in Stereoscopic HTDs", "doi": null, "abstractUrl": "/journal/tg/2002/02/v0129/13rRUwbs2aS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040581", "title": "The Right View from the Wrong Location: Depth Perception in Stereoscopic Multi-User Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040581/13rRUxASuGh", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNySXF2p", "title": "Computer Graphics International Conference", "acronym": "cgi", "groupId": "1000132", "volume": "0", "displayVolume": "0", "year": "1999", "__typename": "ProceedingType" }, "article": { "id": "12OmNx3q6XQ", "doi": "10.1109/CGI.1999.777919", "title": "A Technique for Precise Depth Representation in Stereoscopic Display", "normalizedTitle": "A Technique for Precise Depth Representation in Stereoscopic Display", "abstract": "In observing a 3D virtual object displayed stereoscopically on a large-sized screen, there often exists a difference between the calculated depth and the perceived one of the object. This paper presents a method for reducing such differences of depth. This is performed by modification of both the viewing position and the screen position in stereoscopic calculation. The optimal amount of modification was decided using sample values of depth differences. Effectiveness of the proposed method is discussed with the experimental results.", "abstracts": [ { "abstractType": "Regular", "content": "In observing a 3D virtual object displayed stereoscopically on a large-sized screen, there often exists a difference between the calculated depth and the perceived one of the object. This paper presents a method for reducing such differences of depth. This is performed by modification of both the viewing position and the screen position in stereoscopic calculation. The optimal amount of modification was decided using sample values of depth differences. Effectiveness of the proposed method is discussed with the experimental results.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In observing a 3D virtual object displayed stereoscopically on a large-sized screen, there often exists a difference between the calculated depth and the perceived one of the object. This paper presents a method for reducing such differences of depth. This is performed by modification of both the viewing position and the screen position in stereoscopic calculation. The optimal amount of modification was decided using sample values of depth differences. Effectiveness of the proposed method is discussed with the experimental results.", "fno": "01850080", "keywords": [ "Stereoscopic Display", "Depth Perception", "Virtual Reality", "Pupil Distance" ], "authors": [ { "affiliation": "Nagoya University", "fullName": "Shunsuke Yoshida", "givenName": "Shunsuke", "surname": "Yoshida", "__typename": "ArticleAuthorType" }, { "affiliation": "Nagoya University", "fullName": "Takami Yasuda", "givenName": "Takami", "surname": "Yasuda", "__typename": "ArticleAuthorType" }, { "affiliation": "Nagoya University", "fullName": "Shigeki Yokoi", "givenName": "Shigeki", "surname": "Yokoi", "__typename": "ArticleAuthorType" }, { "affiliation": "Chukyo University", "fullName": "Shin-ya Miyazaki", "givenName": "Shin-ya", "surname": "Miyazaki", "__typename": "ArticleAuthorType" }, { "affiliation": "Chukyo University", "fullName": "Jun-ichi Hasegawa", "givenName": "Jun-ichi", "surname": "Hasegawa", "__typename": "ArticleAuthorType" }, { "affiliation": "Toyota Motor Corporation", "fullName": "Toshihito Hoshino", "givenName": "Toshihito", "surname": "Hoshino", "__typename": "ArticleAuthorType" }, { "affiliation": "Toyota Motor Corporation", "fullName": "Toru Ozeki", "givenName": "Toru", "surname": "Ozeki", "__typename": "ArticleAuthorType" } ], "idPrefix": "cgi", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1999-06-01T00:00:00", "pubType": "proceedings", "pages": "80", "year": "1999", "issn": "1530-1052", "isbn": "0-7695-0185-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "01850078", "articleId": "12OmNCmGNUU", "__typename": "AdjacentArticleType" }, "next": { "fno": "01850085", "articleId": "12OmNCwladD", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwE9OtJ", "title": "2009 Third IEEE International Conference on Space Mission Challenges for Information Technology (SMC-IT 2009)", "acronym": "smc-it", "groupId": "1002093", "volume": "0", "displayVolume": "0", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNxFaLz6", "doi": "10.1109/SMC-IT.2009.24", "title": "Motor Activity-Perception Based Approach for Improving Teleoperation Systems", "normalizedTitle": "Motor Activity-Perception Based Approach for Improving Teleoperation Systems", "abstract": "This paper reports on studies we are conducting to evaluate 3D stereoscopic rendering systems accuracy and efficiency. These studies are part of our project dealing with tele-operation, namely the use of virtual reality technology to increase immersion feeling of teleoperators and to test specific conditions by changing the feeled physical laws. We concentrate on depth perception because it has been shown to be a major factor for simple motor actions like navigation tasks or for complex and dexterous manipulation tasks. We present our stereoscopic rendering system and the methodology we developed to assess it and validate its efficiency in delivering the right stimulations to perceive accurately virtual worlds in 3D. We discuss also its potential use in worlds where physics (gravity forces, viscosity, etc.) aren't conventional.", "abstracts": [ { "abstractType": "Regular", "content": "This paper reports on studies we are conducting to evaluate 3D stereoscopic rendering systems accuracy and efficiency. These studies are part of our project dealing with tele-operation, namely the use of virtual reality technology to increase immersion feeling of teleoperators and to test specific conditions by changing the feeled physical laws. We concentrate on depth perception because it has been shown to be a major factor for simple motor actions like navigation tasks or for complex and dexterous manipulation tasks. We present our stereoscopic rendering system and the methodology we developed to assess it and validate its efficiency in delivering the right stimulations to perceive accurately virtual worlds in 3D. We discuss also its potential use in worlds where physics (gravity forces, viscosity, etc.) aren't conventional.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper reports on studies we are conducting to evaluate 3D stereoscopic rendering systems accuracy and efficiency. These studies are part of our project dealing with tele-operation, namely the use of virtual reality technology to increase immersion feeling of teleoperators and to test specific conditions by changing the feeled physical laws. We concentrate on depth perception because it has been shown to be a major factor for simple motor actions like navigation tasks or for complex and dexterous manipulation tasks. We present our stereoscopic rendering system and the methodology we developed to assess it and validate its efficiency in delivering the right stimulations to perceive accurately virtual worlds in 3D. We discuss also its potential use in worlds where physics (gravity forces, viscosity, etc.) aren't conventional.", "fno": "3637a143", "keywords": [ "Stereoscopic Teleoperation Depth Perception Virtual Reality Virtual Environments" ], "authors": [ { "affiliation": null, "fullName": "Ryad Chellali", "givenName": "Ryad", "surname": "Chellali", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Fabien Dionnet", "givenName": "Fabien", "surname": "Dionnet", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Abdeldjallil Naceri", "givenName": "Abdeldjallil", "surname": "Naceri", "__typename": "ArticleAuthorType" } ], "idPrefix": "smc-it", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-07-01T00:00:00", "pubType": "proceedings", "pages": "143-149", "year": "2009", "issn": null, "isbn": "978-0-7695-3637-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3637a257", "articleId": "12OmNy50gk4", "__typename": "AdjacentArticleType" }, "next": { "fno": "3637a150", "articleId": "12OmNscOUi4", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cw/2008/3381/0/3381a102", "title": "Virtual and Augmented Reality with Head-Tracking for Efficient Teleoperation of Groups of Robots", "doi": null, "abstractUrl": "/proceedings-article/cw/2008/3381a102/12OmNAndinH", "parentPublication": { "id": "proceedings/cw/2008/3381/0", "title": "2008 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/avss/2011/0844/0/06027293", "title": "Stereoscopic viewing facilitates the perception of crowds", "doi": null, "abstractUrl": "/proceedings-article/avss/2011/06027293/12OmNBEpnya", "parentPublication": { "id": "proceedings/avss/2011/0844/0", "title": "2011 8th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/achi/2009/3529/0/3529a310", "title": "Virtual Environments and Scenario Languages for Advanced Teleoperation of Groups of Real Robots: Real Case Application", "doi": null, "abstractUrl": "/proceedings-article/achi/2009/3529a310/12OmNBPtJAZ", "parentPublication": { "id": "proceedings/achi/2009/3529/0", "title": "International Conference on Advances in Computer-Human Interaction", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1993/1363/0/00380802", "title": "The effect of interocular distance upon operator performance using stereoscopic displays to perform virtual depth tasks", "doi": null, "abstractUrl": "/proceedings-article/vrais/1993/00380802/12OmNrJAdLH", "parentPublication": { "id": "proceedings/vrais/1993/1363/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fgcns/2008/3546/3/3546c031", "title": "A Novel Approach to Depth Image Based Rendering Based on Non-Uniform Scaling of Depth Values", "doi": null, "abstractUrl": "/proceedings-article/fgcns/2008/3546c031/12OmNwJybNz", "parentPublication": { "id": "proceedings/fgcns/2008/3546/3", "title": "Future Generation Communication and Networking Symposia, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icycs/2008/3398/0/3398a831", "title": "Precise Depth Perception in Projective Stereoscopic Display", "doi": null, "abstractUrl": "/proceedings-article/icycs/2008/3398a831/12OmNwtEEM5", "parentPublication": { "id": "proceedings/icycs/2008/3398/0", "title": "2008 9th International Conference for Young Computer Scientists", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/1999/0185/0/01850080", "title": "A Technique for Precise Depth Representation in Stereoscopic Display", "doi": null, "abstractUrl": "/proceedings-article/cgi/1999/01850080/12OmNx3q6XQ", "parentPublication": { "id": "proceedings/cgi/1999/0185/0", "title": "Computer Graphics International Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/apwcs/2010/4003/0/4003a147", "title": "A New Virtual View Rendering Method Based on Depth Image", "doi": null, "abstractUrl": "/proceedings-article/apwcs/2010/4003a147/12OmNy6HQX6", "parentPublication": { "id": "proceedings/apwcs/2010/4003/0", "title": "Wearable Computing Systems, Asia-Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040581", "title": "The Right View from the Wrong Location: Depth Perception in Stereoscopic Multi-User Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040581/13rRUxASuGh", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2004/01/v0015", "title": "Interactive Stereoscopic Rendering of Volumetric Environments", "doi": null, "abstractUrl": "/journal/tg/2004/01/v0015/13rRUyYBlgp", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNs0C9QC", "title": "2016 International Symposium on Computer, Consumer and Control (IS3C)", "acronym": "is3c", "groupId": "1801670", "volume": "0", "displayVolume": "0", "year": "2016", "__typename": "ProceedingType" }, "article": { "id": "12OmNzUPpta", "doi": "10.1109/IS3C.2016.81", "title": "A Study of Effects of Perceptual Cues on Presence for the Elderly in 3D Virtual Store", "normalizedTitle": "A Study of Effects of Perceptual Cues on Presence for the Elderly in 3D Virtual Store", "abstract": "Virtual environments (VEs) technology applied in web shops provided a 3D perspective to customers for more real sense on goods and shopping environment. A sense of presence is one of the critical components required by any effective VEs. However, previous studies on cognitive aging have found that certain aspects of human information-processing abilities are negatively correlated with age. Therefore, when the quality of depth perceptual cues is good, whether the perceptual confliction for the elderly will be less to influence the feeling of presence or not? An experiment addressed presence, and performance of 3D virtual store in the elderly participation with auto-stereoscopic, stereoscopic and monocular display in good/poor depth perceptual cues. Results showed that 3D virtual store via auto-stereoscopic display with high quality depth perceptual cues will produce good sense and realism in stereopsis. However, the response time of objects-searching with 3D displays was longer than monocular display.", "abstracts": [ { "abstractType": "Regular", "content": "Virtual environments (VEs) technology applied in web shops provided a 3D perspective to customers for more real sense on goods and shopping environment. A sense of presence is one of the critical components required by any effective VEs. However, previous studies on cognitive aging have found that certain aspects of human information-processing abilities are negatively correlated with age. Therefore, when the quality of depth perceptual cues is good, whether the perceptual confliction for the elderly will be less to influence the feeling of presence or not? An experiment addressed presence, and performance of 3D virtual store in the elderly participation with auto-stereoscopic, stereoscopic and monocular display in good/poor depth perceptual cues. Results showed that 3D virtual store via auto-stereoscopic display with high quality depth perceptual cues will produce good sense and realism in stereopsis. However, the response time of objects-searching with 3D displays was longer than monocular display.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Virtual environments (VEs) technology applied in web shops provided a 3D perspective to customers for more real sense on goods and shopping environment. A sense of presence is one of the critical components required by any effective VEs. However, previous studies on cognitive aging have found that certain aspects of human information-processing abilities are negatively correlated with age. Therefore, when the quality of depth perceptual cues is good, whether the perceptual confliction for the elderly will be less to influence the feeling of presence or not? An experiment addressed presence, and performance of 3D virtual store in the elderly participation with auto-stereoscopic, stereoscopic and monocular display in good/poor depth perceptual cues. Results showed that 3D virtual store via auto-stereoscopic display with high quality depth perceptual cues will produce good sense and realism in stereopsis. However, the response time of objects-searching with 3D displays was longer than monocular display.", "fno": "3071a283", "keywords": [ "Three Dimensional Displays", "Two Dimensional Displays", "Stereo Image Processing", "Senior Citizens", "Glass", "Time Factors", "Search Problems", "Stereoscopic Display", "3 D Virtual Store", "Elderly", "Presence", "Depth Perceptual Cues" ], "authors": [ { "affiliation": null, "fullName": "Cheng-Li Liu", "givenName": "Cheng-Li", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Shiaw-Tsyr Uang", "givenName": "Shiaw-Tsyr", "surname": "Uang", "__typename": "ArticleAuthorType" } ], "idPrefix": "is3c", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2016-07-01T00:00:00", "pubType": "proceedings", "pages": "283-286", "year": "2016", "issn": null, "isbn": "978-1-5090-3071-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3071a279", "articleId": "12OmNqEji0d", "__typename": "AdjacentArticleType" }, "next": { "fno": "3071a287", "articleId": "12OmNBOll7H", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2017/6647/0/07892274", "title": "Mechanism of integrating force and vibrotactile cues for 3D user interaction within virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892274/12OmNqH9hid", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/percomw/2016/1941/0/07457134", "title": "Twinkle megane: Near-eye LED indicators on glasses in tele-guidance for elderly", "doi": null, "abstractUrl": "/proceedings-article/percomw/2016/07457134/12OmNwGZNPu", "parentPublication": { "id": "proceedings/percomw/2016/1941/0", "title": "2016 IEEE International Conference on Pervasive Computing and Communication Workshops (PerCom Workshops)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccsit/2009/4519/0/05234642", "title": "A hierarchy grouping model based on gestalt perceptual cues", "doi": null, "abstractUrl": "/proceedings-article/iccsit/2009/05234642/12OmNxHJ9uB", "parentPublication": { "id": "proceedings/iccsit/2009/4519/0", "title": "Computer Science and Information Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dvis/2014/6826/0/07160093", "title": "The (possible) utility of stereoscopic 3D displays for information visualization: The good, the bad, and the ugly", "doi": null, "abstractUrl": "/proceedings-article/3dvis/2014/07160093/12OmNxWcHag", "parentPublication": { "id": "proceedings/3dvis/2014/6826/0", "title": "2014 IEEE VIS International Workshop on 3DVis (3DVis)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1995/7084/0/70840074", "title": "Presence in virtual environments as a function of visual and auditory cues", "doi": null, "abstractUrl": "/proceedings-article/vrais/1995/70840074/12OmNzlUKP6", "parentPublication": { "id": "proceedings/vrais/1995/7084/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446424", "title": "An Exploration on the Integration of Vibrotactile and Force Cues for 3D Interactive Tasks", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446424/13bd1AITnaI", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2014/03/ttg2014030391", "title": "An Evaluation of Depth Enhancing Perceptual Cues for Vascular Volume Visualization in Neurosurgery", "doi": null, "abstractUrl": "/journal/tg/2014/03/ttg2014030391/13rRUxjQybS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/percom-workshops/2019/9151/0/08730829", "title": "Ambient Cues of Kitchen Counter in Guiding Cooking Activities for Alzheimer&#x0027;s Patient", "doi": null, "abstractUrl": "/proceedings-article/percom-workshops/2019/08730829/1aDSFCMozYI", "parentPublication": { "id": "proceedings/percom-workshops/2019/9151/0", "title": "2019 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2019/5434/0/543400a084", "title": "User Interface Evaluation Methods for Elderly: A Systematic Review", "doi": null, "abstractUrl": "/proceedings-article/svr/2019/543400a084/1fHjyvPAmuQ", "parentPublication": { "id": "proceedings/svr/2019/5434/0", "title": "2019 21st Symposium on Virtual and Augmented Reality (SVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/11/09523847", "title": "Multifocal Stereoscopic Projection Mapping", "doi": null, "abstractUrl": "/journal/tg/2021/11/09523847/1wpqmNfLX9e", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1gRxdWFgE3m", "title": "2019 10th International Conference on Information Technology in Medicine and Education (ITME)", "acronym": "itme", "groupId": "1002567", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1gRxnxuyvDy", "doi": "10.1109/ITME.2019.00173", "title": "The Monocular Stereoscopic Model Based on an Ordered Wave Particle Swarm", "normalizedTitle": "The Monocular Stereoscopic Model Based on an Ordered Wave Particle Swarm", "abstract": "Visual neuroscience data support monocular depth cues such as texture gradients. Although the binocular disparity energy model has been developed, it concentrates on binocular depth perception. We present a monocular stereoscopic model based on an ordered wave particle swarm to explain monocular cues to depth. The ordering of the particle swarm is triggered by the stereoscopic potential embedded in textures to create monocular depth perception. First, each receptive field (RF) as a wave particle was expressed as a wave equation, and a stereoscopic potential field was extracted from monocular depth cues as the initial conditions. Second, the solutions of wave equations showed that the wave particles affected by the potential field were ordered by frequency. Finally, the consistency of the monocular model with the Gabor filter was discussed.", "abstracts": [ { "abstractType": "Regular", "content": "Visual neuroscience data support monocular depth cues such as texture gradients. Although the binocular disparity energy model has been developed, it concentrates on binocular depth perception. We present a monocular stereoscopic model based on an ordered wave particle swarm to explain monocular cues to depth. The ordering of the particle swarm is triggered by the stereoscopic potential embedded in textures to create monocular depth perception. First, each receptive field (RF) as a wave particle was expressed as a wave equation, and a stereoscopic potential field was extracted from monocular depth cues as the initial conditions. Second, the solutions of wave equations showed that the wave particles affected by the potential field were ordered by frequency. Finally, the consistency of the monocular model with the Gabor filter was discussed.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Visual neuroscience data support monocular depth cues such as texture gradients. Although the binocular disparity energy model has been developed, it concentrates on binocular depth perception. We present a monocular stereoscopic model based on an ordered wave particle swarm to explain monocular cues to depth. The ordering of the particle swarm is triggered by the stereoscopic potential embedded in textures to create monocular depth perception. First, each receptive field (RF) as a wave particle was expressed as a wave equation, and a stereoscopic potential field was extracted from monocular depth cues as the initial conditions. Second, the solutions of wave equations showed that the wave particles affected by the potential field were ordered by frequency. Finally, the consistency of the monocular model with the Gabor filter was discussed.", "fno": "391800a758", "keywords": [ "Gabor Filters", "Neurophysiology", "Particle Swarm Optimisation", "Stereo Image Processing", "Three Dimensional Displays", "Visual Perception", "Binocular Depth Perception", "Monocular Stereoscopic Model", "Ordered Wave Particle Swarm", "Monocular Depth Cues", "Monocular Depth Perception", "Wave Equation", "Stereoscopic Potential Field", "Texture Gradients", "Binocular Disparity Energy Model", "Visual Neuroscience Data", "Receptive Field", "Gabor Filter", "Stereo Image Processing", "Retina", "Radio Frequency", "Particle Swarm Optimization", "Propagation", "Gabor Filters", "Mathematical Model", "Monocular Stereoscopic Cues", "Wave Particles Swarm", "Receptive Fields", "Stereoscopic Potential Fields", "The Gabor Filter" ], "authors": [ { "affiliation": "Shandong Normal University", "fullName": "Jihua Wang", "givenName": "Jihua", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "Shandong Normal University", "fullName": "Huanchun Yuan", "givenName": "Huanchun", "surname": "Yuan", "__typename": "ArticleAuthorType" }, { "affiliation": "Shandong Normal University", "fullName": "Huayu Wang", "givenName": "Huayu", "surname": "Wang", "__typename": "ArticleAuthorType" } ], "idPrefix": "itme", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-08-01T00:00:00", "pubType": "proceedings", "pages": "758-762", "year": "2019", "issn": null, "isbn": "978-1-7281-3918-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "391800a753", "articleId": "1gRxkM2gKpq", "__typename": "AdjacentArticleType" }, "next": { "fno": "391800a763", "articleId": "1gRxkPkmkVO", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icme/2017/6067/0/08019337", "title": "Reduced reference stereoscopic image quality assessment based on entropy of classified primitives", "doi": null, "abstractUrl": "/proceedings-article/icme/2017/08019337/12OmNAndisf", "parentPublication": { "id": "proceedings/icme/2017/6067/0", "title": "2017 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/lcn/2012/1565/0/06423586", "title": "3DSVAT: A 3D Stereoscopic Vulnerability Assessment Tool for network security", "doi": null, "abstractUrl": "/proceedings-article/lcn/2012/06423586/12OmNqzcvC0", "parentPublication": { "id": "proceedings/lcn/2012/1565/0", "title": "37th Annual IEEE Conference on Local Computer Networks", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2013/2840/0/2840a073", "title": "Joint Subspace Stabilization for Stereoscopic Video", "doi": null, "abstractUrl": "/proceedings-article/iccv/2013/2840a073/12OmNxT56Af", "parentPublication": { "id": "proceedings/iccv/2013/2840/0", "title": "2013 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2012/4875/0/4875a306", "title": "Towards Automatic Stereoscopic Video Synthesis from a Casual Monocular Video", "doi": null, "abstractUrl": "/proceedings-article/ism/2012/4875a306/12OmNyKrHdW", "parentPublication": { "id": "proceedings/ism/2012/4875/0", "title": "2012 IEEE International Symposium on Multimedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/is3c/2016/3071/0/3071a283", "title": "A Study of Effects of Perceptual Cues on Presence for the Elderly in 3D Virtual Store", "doi": null, "abstractUrl": "/proceedings-article/is3c/2016/3071a283/12OmNzUPpta", "parentPublication": { "id": "proceedings/is3c/2016/3071/0", "title": "2016 International Symposium on Computer, Consumer and Control (IS3C)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/04/v0686", "title": "Stereoscopic Video Synthesis from a Monocular Video", "doi": null, "abstractUrl": "/journal/tg/2007/04/v0686/13rRUwcAqq7", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040581", "title": "The Right View from the Wrong Location: Depth Perception in Stereoscopic Multi-User Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040581/13rRUxASuGh", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/05/09714042", "title": "Stereopsis Only: Validation of a Monocular Depth Cues Reduced Gamified Virtual Reality with Reaction Time Measurement", "doi": null, "abstractUrl": "/journal/tg/2022/05/09714042/1B0XZrABGk8", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300j741", "title": "Learning to Explore Intrinsic Saliency for Stereoscopic Video", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300j741/1gyrLoWsF1u", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/11/09523847", "title": "Multifocal Stereoscopic Projection Mapping", "doi": null, "abstractUrl": "/journal/tg/2021/11/09523847/1wpqmNfLX9e", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAYXWAF", "title": "2016 IEEE Virtual Reality (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2016", "__typename": "ProceedingType" }, "article": { "id": "12OmNAnMuMd", "doi": "10.1109/VR.2016.7504693", "title": "A calibration method for optical see-through head-mounted displays with a depth camera", "normalizedTitle": "A calibration method for optical see-through head-mounted displays with a depth camera", "abstract": "We propose a fast and accurate calibration method for the optical see-through (OST) head-mounted displays (HMD), taking advantage of a low-cost time-of-flight depth-camera. Recently, affordable OST-HMDs and depth-cameras are widely appearing in the commercial market. In order to correctly reflect the user experience into the calibration process, our method demands a user wearing the HMD to repeatedly point at rendered virtual circles with their fingertips. From the repeated calibration data, we perform two stages of full calibration and simplified calibration, to compute key calibration parameters. The full calibration is required when the depth-camera is first installed to the HMD, and afterwards only the simplified calibration is performed whenever a user wears it again. Our experimental results show that the full and simplified calibration can be achieved with 10 and 5 user's repetitions (theoretically 3 and 2 at minimum), which are significantly less than about 20 of the stereo-SPAAM, one of the most popular existing calibration techniques. We also demonstrate that the 3D position errors of our calibration become much quickly smaller than those of the state-of-the-art method.", "abstracts": [ { "abstractType": "Regular", "content": "We propose a fast and accurate calibration method for the optical see-through (OST) head-mounted displays (HMD), taking advantage of a low-cost time-of-flight depth-camera. Recently, affordable OST-HMDs and depth-cameras are widely appearing in the commercial market. In order to correctly reflect the user experience into the calibration process, our method demands a user wearing the HMD to repeatedly point at rendered virtual circles with their fingertips. From the repeated calibration data, we perform two stages of full calibration and simplified calibration, to compute key calibration parameters. The full calibration is required when the depth-camera is first installed to the HMD, and afterwards only the simplified calibration is performed whenever a user wears it again. Our experimental results show that the full and simplified calibration can be achieved with 10 and 5 user's repetitions (theoretically 3 and 2 at minimum), which are significantly less than about 20 of the stereo-SPAAM, one of the most popular existing calibration techniques. We also demonstrate that the 3D position errors of our calibration become much quickly smaller than those of the state-of-the-art method.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose a fast and accurate calibration method for the optical see-through (OST) head-mounted displays (HMD), taking advantage of a low-cost time-of-flight depth-camera. Recently, affordable OST-HMDs and depth-cameras are widely appearing in the commercial market. In order to correctly reflect the user experience into the calibration process, our method demands a user wearing the HMD to repeatedly point at rendered virtual circles with their fingertips. From the repeated calibration data, we perform two stages of full calibration and simplified calibration, to compute key calibration parameters. The full calibration is required when the depth-camera is first installed to the HMD, and afterwards only the simplified calibration is performed whenever a user wears it again. Our experimental results show that the full and simplified calibration can be achieved with 10 and 5 user's repetitions (theoretically 3 and 2 at minimum), which are significantly less than about 20 of the stereo-SPAAM, one of the most popular existing calibration techniques. We also demonstrate that the 3D position errors of our calibration become much quickly smaller than those of the state-of-the-art method.", "fno": "07504693", "keywords": [ "Calibration", "Cameras", "Three Dimensional Displays", "Adaptive Optics", "Head", "Magnetic Heads", "Mathematical Model", "Depth Camera", "Calibration", "Optical See Through Head Mounted Display" ], "authors": [ { "affiliation": "Seoul National University", "fullName": "Hanseul Jun", "givenName": "Hanseul", "surname": "Jun", "__typename": "ArticleAuthorType" }, { "affiliation": "Seoul National University", "fullName": "Gunhee Kim", "givenName": "Gunhee", "surname": "Kim", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2016-03-01T00:00:00", "pubType": "proceedings", "pages": "103-111", "year": "2016", "issn": "2375-5334", "isbn": "978-1-5090-0836-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07504692", "articleId": "12OmNzRZpYz", "__typename": "AdjacentArticleType" }, "next": { "fno": "07504694", "articleId": "12OmNzWx00N", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cw/2017/2089/0/2089a033", "title": "User Friendly Calibration for Tracking of Optical Stereo See-Through Head Worn Displays for Augmented Reality", "doi": null, "abstractUrl": "/proceedings-article/cw/2017/2089a033/12OmNAfPISE", "parentPublication": { "id": "proceedings/cw/2017/2089/0", "title": "2017 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948513", "title": "Google glass, The META and Co. How to calibrate optical see-through head mounted displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948513/12OmNB8TUim", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2015/7660/0/7660a043", "title": "Simultaneous Direct and Augmented View Distortion Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2015/7660a043/12OmNC1oT64", "parentPublication": { "id": "proceedings/ismar/2015/7660/0", "title": "2015 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2015/1727/0/07223450", "title": "Evaluating optical see-through head-mounted display calibration via frustum visualization", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223450/12OmNrAv3Ap", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948424", "title": "Performance and sensitivity analysis of INDICA: INteraction-Free DIsplay CAlibration for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948424/12OmNyYm2oO", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/09/08052554", "title": "A Survey of Calibration Methods for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2018/09/08052554/13rRUILtJqY", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07021939", "title": "Subjective Evaluation of a Semi-Automatic Optical See-Through Head-Mounted Display Calibration Technique", "doi": null, "abstractUrl": "/journal/tg/2015/04/07021939/13rRUwInvyB", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07064856", "title": "Light-Field Correction for Spatial Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/04/07064856/13rRUwjGoG5", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/11/07523375", "title": "Gaussian Light Field: Estimation of Viewpoint-Dependent Blur for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2016/11/07523375/13rRUxYINfi", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090625", "title": "Automatic Calibration of Commercial Optical See-Through Head-Mounted Displays for Medical Applications", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090625/1jIxwp2g0VO", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqJ8taA", "title": "Augmented Reality, International Workshop on", "acronym": "iwar", "groupId": "1000063", "volume": "0", "displayVolume": "0", "year": "1999", "__typename": "ProceedingType" }, "article": { "id": "12OmNBcAGLe", "doi": "10.1109/IWAR.1999.803809", "title": "Marker Tracking and HMD Calibration for a Video-Based Augmented Reality Conferencing System", "normalizedTitle": "Marker Tracking and HMD Calibration for a Video-Based Augmented Reality Conferencing System", "abstract": "We describe an augmented reality conferencing system which uses the overlay of virtual images on the real world. Remote collaborators are represented on Virtual Monitors which can be freely positioned about a user in space. Users can collaboratively view and interact with virtual objects using a shared virtual whiteboard. This is possible through precise virtual image registration using fast and accurate computer vision techniques and HMD calibration. We propose a method for tracking fiducial markers and a calibration method for optical see-through HMD based on the marker tracking.", "abstracts": [ { "abstractType": "Regular", "content": "We describe an augmented reality conferencing system which uses the overlay of virtual images on the real world. Remote collaborators are represented on Virtual Monitors which can be freely positioned about a user in space. Users can collaboratively view and interact with virtual objects using a shared virtual whiteboard. This is possible through precise virtual image registration using fast and accurate computer vision techniques and HMD calibration. We propose a method for tracking fiducial markers and a calibration method for optical see-through HMD based on the marker tracking.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We describe an augmented reality conferencing system which uses the overlay of virtual images on the real world. Remote collaborators are represented on Virtual Monitors which can be freely positioned about a user in space. Users can collaboratively view and interact with virtual objects using a shared virtual whiteboard. This is possible through precise virtual image registration using fast and accurate computer vision techniques and HMD calibration. We propose a method for tracking fiducial markers and a calibration method for optical see-through HMD based on the marker tracking.", "fno": "03590085", "keywords": [ "Augmented Reality", "Teleconferencing", "CSCW", "Optical See Through AR", "HMD Calibration" ], "authors": [ { "affiliation": "Hiroshima City University", "fullName": "Hirokazu Kato", "givenName": "Hirokazu", "surname": "Kato", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Washington", "fullName": "Mark Billinghurst", "givenName": "Mark", "surname": "Billinghurst", "__typename": "ArticleAuthorType" } ], "idPrefix": "iwar", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1999-10-01T00:00:00", "pubType": "proceedings", "pages": "85", "year": "1999", "issn": null, "isbn": "0-7695-0359-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "03590075", "articleId": "12OmNxTVU20", "__typename": "AdjacentArticleType" }, "next": { "fno": "03590095", "articleId": "12OmNyQGRY2", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAsk4yu", "title": "The Second IEEE and ACM International Symposium on Mixed and Augmented Reality, 2003. Proceedings.", "acronym": "ismar", "groupId": "1000465", "volume": "0", "displayVolume": "0", "year": "2003", "__typename": "ProceedingType" }, "article": { "id": "12OmNCeK2b7", "doi": "10.1109/ISMAR.2003.1240699", "title": "Evaluation of Calibration Procedures for Optical See-Through Head-Mounted Displays", "normalizedTitle": "Evaluation of Calibration Procedures for Optical See-Through Head-Mounted Displays", "abstract": "Optical see-through head-mounted displays (HMDs) are less commonly used because they are difficult to accurately calibrate. In this article, we report a user study to compare the accuracy of 4 variants of the SPAAM calibration method. Among the 4 variants, Stylus-marker calibration, where the user aligns a crosshair projected in the HMD with a tracked stylus tip, achieved the most accurate result. A decomposition and analysis of the calibration matrices from the trials is performed and the characteristics of the computed calibration matrices are examined. A physiological engineering point of view is also discussed to explain why calibrating optical see-through HMD is so difficult for users.", "abstracts": [ { "abstractType": "Regular", "content": "Optical see-through head-mounted displays (HMDs) are less commonly used because they are difficult to accurately calibrate. In this article, we report a user study to compare the accuracy of 4 variants of the SPAAM calibration method. Among the 4 variants, Stylus-marker calibration, where the user aligns a crosshair projected in the HMD with a tracked stylus tip, achieved the most accurate result. A decomposition and analysis of the calibration matrices from the trials is performed and the characteristics of the computed calibration matrices are examined. A physiological engineering point of view is also discussed to explain why calibrating optical see-through HMD is so difficult for users.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Optical see-through head-mounted displays (HMDs) are less commonly used because they are difficult to accurately calibrate. In this article, we report a user study to compare the accuracy of 4 variants of the SPAAM calibration method. Among the 4 variants, Stylus-marker calibration, where the user aligns a crosshair projected in the HMD with a tracked stylus tip, achieved the most accurate result. A decomposition and analysis of the calibration matrices from the trials is performed and the characteristics of the computed calibration matrices are examined. A physiological engineering point of view is also discussed to explain why calibrating optical see-through HMD is so difficult for users.", "fno": "20060161", "keywords": [], "authors": [ { "affiliation": "Michigan State University, East Lansing", "fullName": "Arthur Tang", "givenName": "Arthur", "surname": "Tang", "__typename": "ArticleAuthorType" }, { "affiliation": "Michigan State University, East Lansing", "fullName": "Ji Zhou", "givenName": "Ji", "surname": "Zhou", "__typename": "ArticleAuthorType" }, { "affiliation": "Michigan State University, East Lansing", "fullName": "Charles Owen", "givenName": "Charles", "surname": "Owen", "__typename": "ArticleAuthorType" } ], "idPrefix": "ismar", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2003-10-01T00:00:00", "pubType": "proceedings", "pages": "161", "year": "2003", "issn": null, "isbn": "0-7695-2006-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "20060151", "articleId": "12OmNBhpS86", "__typename": "AdjacentArticleType" }, "next": { "fno": "20060169", "articleId": "12OmNxyDZdG", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2016/0836/0/07504693", "title": "A calibration method for optical see-through head-mounted displays with a depth camera", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504693/12OmNAnMuMd", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948513", "title": "Google glass, The META and Co. How to calibrate optical see-through head mounted displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948513/12OmNB8TUim", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2015/7660/0/7660a043", "title": "Simultaneous Direct and Augmented View Distortion Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2015/7660a043/12OmNC1oT64", "parentPublication": { "id": "proceedings/ismar/2015/7660/0", "title": "2015 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2014/3624/0/06798846", "title": "Interaction-free calibration for optical see-through head-mounted displays based on 3D Eye localization", "doi": null, "abstractUrl": "/proceedings-article/3dui/2014/06798846/12OmNCdBDWL", "parentPublication": { "id": "proceedings/3dui/2014/3624/0", "title": "2014 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2016/0836/0/07504739", "title": "Evaluation of hand and stylus based calibration for optical see-through head-mounted displays using leap motion", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504739/12OmNxE2mZD", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948481", "title": "[DEMO] INDICA : Interaction-free display calibration for optical see-through head-mounted displays based on 3D eye localization", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948481/12OmNy1SFEx", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/11/07165643", "title": "Semi-Parametric Color Reproduction Method for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/11/07165643/13rRUILtJzB", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07064856", "title": "Light-Field Correction for Spatial Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/04/07064856/13rRUwjGoG5", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07012105", "title": "Corneal-Imaging Calibration for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/04/07012105/13rRUxE04tC", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090625", "title": "Automatic Calibration of Commercial Optical See-Through Head-Mounted Displays for Medical Applications", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090625/1jIxwp2g0VO", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvAiSpZ", "title": "2015 IEEE Virtual Reality (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNrAv3Ap", "doi": "10.1109/VR.2015.7223450", "title": "Evaluating optical see-through head-mounted display calibration via frustum visualization", "normalizedTitle": "Evaluating optical see-through head-mounted display calibration via frustum visualization", "abstract": "Effectively evaluating optical see-through (OST) head-mounted display (HMD) calibration is problematic and largely relies on feedback from the user. Studies evaluating OST HMD calibration, such as those by McGarrity, Tang, and Navab et al. [2, 3, 1], utilize user interaction methods, such as touch pads, to facilitate on-line evaluation and correction of calibration results. In all of these studies, however, only the users themselves receive any visual feedback related to the calibration quality or the corrective actions taken to improve it. In this video, we present the use of standard frustum visualization to provide calibration quality information to the researcher in real time. We use a standard Single Point Active Alignment Method (SPAAM) calibration, [4], after which both the eye location estimate and resulting intrinsic values are displayed superimposed onto the user. Presenting the eye position relative to the user's head benefits studies on system error sources, and rendering on-screen visuals also allows outside observers to identify calibration issues and offer corrective suggestions. We believe that techniques, such as frustum visualization, will expand the amount of information available for evaluating calibration results, and will greatly aid those investigating new and improved calibration procedures.", "abstracts": [ { "abstractType": "Regular", "content": "Effectively evaluating optical see-through (OST) head-mounted display (HMD) calibration is problematic and largely relies on feedback from the user. Studies evaluating OST HMD calibration, such as those by McGarrity, Tang, and Navab et al. [2, 3, 1], utilize user interaction methods, such as touch pads, to facilitate on-line evaluation and correction of calibration results. In all of these studies, however, only the users themselves receive any visual feedback related to the calibration quality or the corrective actions taken to improve it. In this video, we present the use of standard frustum visualization to provide calibration quality information to the researcher in real time. We use a standard Single Point Active Alignment Method (SPAAM) calibration, [4], after which both the eye location estimate and resulting intrinsic values are displayed superimposed onto the user. Presenting the eye position relative to the user's head benefits studies on system error sources, and rendering on-screen visuals also allows outside observers to identify calibration issues and offer corrective suggestions. We believe that techniques, such as frustum visualization, will expand the amount of information available for evaluating calibration results, and will greatly aid those investigating new and improved calibration procedures.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Effectively evaluating optical see-through (OST) head-mounted display (HMD) calibration is problematic and largely relies on feedback from the user. Studies evaluating OST HMD calibration, such as those by McGarrity, Tang, and Navab et al. [2, 3, 1], utilize user interaction methods, such as touch pads, to facilitate on-line evaluation and correction of calibration results. In all of these studies, however, only the users themselves receive any visual feedback related to the calibration quality or the corrective actions taken to improve it. In this video, we present the use of standard frustum visualization to provide calibration quality information to the researcher in real time. We use a standard Single Point Active Alignment Method (SPAAM) calibration, [4], after which both the eye location estimate and resulting intrinsic values are displayed superimposed onto the user. Presenting the eye position relative to the user's head benefits studies on system error sources, and rendering on-screen visuals also allows outside observers to identify calibration issues and offer corrective suggestions. We believe that techniques, such as frustum visualization, will expand the amount of information available for evaluating calibration results, and will greatly aid those investigating new and improved calibration procedures.", "fno": "07223450", "keywords": [ "Calibration", "Visualization", "Adaptive Optics", "Optical Feedback", "Augmented Reality", "Optical Imaging", "Standards", "Frustum Visualization", "Calibration", "OST HMD" ], "authors": [ { "affiliation": "Mississippi State University, Department of Computer Science and Engineering", "fullName": "Kenneth R. Moser", "givenName": "Kenneth R.", "surname": "Moser", "__typename": "ArticleAuthorType" }, { "affiliation": "Mississippi State University, Department of Computer Science and Engineering", "fullName": "J. Edward Swan", "givenName": "J. Edward", "surname": "Swan", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-03-01T00:00:00", "pubType": "proceedings", "pages": "371-371", "year": "2015", "issn": null, "isbn": "978-1-4799-1727-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07223449", "articleId": "12OmNyeWdKP", "__typename": "AdjacentArticleType" }, "next": { "fno": "07223451", "articleId": "12OmNzE54CB", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2016/0836/0/07504693", "title": "A calibration method for optical see-through head-mounted displays with a depth camera", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504693/12OmNAnMuMd", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948513", "title": "Google glass, The META and Co. How to calibrate optical see-through head mounted displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948513/12OmNB8TUim", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2012/4660/0/06402574", "title": "Occlusion capable optical see-through head-mounted display using freeform optics", "doi": null, "abstractUrl": "/proceedings-article/ismar/2012/06402574/12OmNBEpnEt", "parentPublication": { "id": "proceedings/ismar/2012/4660/0", "title": "2012 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2015/7660/0/7660a043", "title": "Simultaneous Direct and Augmented View Distortion Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2015/7660a043/12OmNC1oT64", "parentPublication": { "id": "proceedings/ismar/2015/7660/0", "title": "2015 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2017/6647/0/07892255", "title": "Robust optical see-through head-mounted display calibration: Taking anisotropic nature of user interaction errors into account", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892255/12OmNxvO04e", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2014/2871/0/06802089", "title": "Quantification of error from system and environmental sources in Optical See-Through head mounted display calibration methods", "doi": null, "abstractUrl": "/proceedings-article/vr/2014/06802089/12OmNxwncbX", "parentPublication": { "id": "proceedings/vr/2014/2871/0", "title": "2014 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948424", "title": "Performance and sensitivity analysis of INDICA: INteraction-Free DIsplay CAlibration for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948424/12OmNyYm2oO", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07021939", "title": "Subjective Evaluation of a Semi-Automatic Optical See-Through Head-Mounted Display Calibration Technique", "doi": null, "abstractUrl": "/journal/tg/2015/04/07021939/13rRUwInvyB", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07064856", "title": "Light-Field Correction for Spatial Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/04/07064856/13rRUwjGoG5", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/11/07523375", "title": "Gaussian Light Field: Estimation of Viewpoint-Dependent Blur for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2016/11/07523375/13rRUxYINfi", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyUWQR6", "title": "Virtual Reality Annual International Symposium", "acronym": "vrais", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "1993", "__typename": "ProceedingType" }, "article": { "id": "12OmNwwuDRm", "doi": "10.1109/VRAIS.1993.380772", "title": "Calibration of head-mounted displays for augmented reality applications", "normalizedTitle": "Calibration of head-mounted displays for augmented reality applications", "abstract": "The authors have developed \"augmented reality\" technology, consisting of a see-through head-mounted display, a robust, accurate position/orientation sensor, and their supporting electronics and software. Their primary goal is to apply this technology to touch labor manufacturing processes, enabling a factory worker to view index markings or instructions as if they were painted on the surface of a workpiece. In order to accurately project graphics onto specific points of a workpiece, it is necessary to have the coordinates of the workpiece, the display's virtual screen, the position sensor, and the user's eyes in the same coordinate system. The linear transformation and projection of each point to be displayed from world coordinates to virtual screen coordinates are described, and the experimental procedures for determining the correct values of the calibration parameters are characterized.", "abstracts": [ { "abstractType": "Regular", "content": "The authors have developed \"augmented reality\" technology, consisting of a see-through head-mounted display, a robust, accurate position/orientation sensor, and their supporting electronics and software. Their primary goal is to apply this technology to touch labor manufacturing processes, enabling a factory worker to view index markings or instructions as if they were painted on the surface of a workpiece. In order to accurately project graphics onto specific points of a workpiece, it is necessary to have the coordinates of the workpiece, the display's virtual screen, the position sensor, and the user's eyes in the same coordinate system. The linear transformation and projection of each point to be displayed from world coordinates to virtual screen coordinates are described, and the experimental procedures for determining the correct values of the calibration parameters are characterized.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The authors have developed \"augmented reality\" technology, consisting of a see-through head-mounted display, a robust, accurate position/orientation sensor, and their supporting electronics and software. Their primary goal is to apply this technology to touch labor manufacturing processes, enabling a factory worker to view index markings or instructions as if they were painted on the surface of a workpiece. In order to accurately project graphics onto specific points of a workpiece, it is necessary to have the coordinates of the workpiece, the display's virtual screen, the position sensor, and the user's eyes in the same coordinate system. The linear transformation and projection of each point to be displayed from world coordinates to virtual screen coordinates are described, and the experimental procedures for determining the correct values of the calibration parameters are characterized.", "fno": "00380772", "keywords": [ "Calibration Parameters", "Augmented Reality Applications", "See Through Head Mounted Display", "Accurate Position Orientation Sensor", "Touch Labor Manufacturing Processes", "Factory Worker", "Index Markings", "Workpiece", "Virtual Screen", "Position Sensor", "Coordinate System", "Linear Transformation", "World Coordinates", "Experimental Procedures" ], "authors": [ { "affiliation": "Boeing Comput. Services, Seattle, WA, USA", "fullName": "A.L. Janin", "givenName": "A.L.", "surname": "Janin", "__typename": "ArticleAuthorType" }, { "affiliation": "Boeing Comput. Services, Seattle, WA, USA", "fullName": "D.W. Mizell", "givenName": "D.W.", "surname": "Mizell", "__typename": "ArticleAuthorType" }, { "affiliation": "Boeing Comput. Services, Seattle, WA, USA", "fullName": "T.P. Caudell", "givenName": "T.P.", "surname": "Caudell", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrais", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1993-09-01T00:00:00", "pubType": "proceedings", "pages": "246-255", "year": "1993", "issn": null, "isbn": "0-7803-1363-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "00380771", "articleId": "12OmNqI04Kn", "__typename": "AdjacentArticleType" }, "next": { "fno": "00380773", "articleId": "12OmNz4SOD4", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwMXnv0", "title": "2014 IEEE Virtual Reality (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNxwENpf", "doi": "10.1109/VR.2014.6802063", "title": "Automated calibration of display characteristics (ACDC) for head-mounted displays and arbitrary surfaces", "normalizedTitle": "Automated calibration of display characteristics (ACDC) for head-mounted displays and arbitrary surfaces", "abstract": "In this document we present a method for calibrating head-mounted displays and other display surfaces using an automated, low-cost camera system. A unique aspect of this method is that the calibration of geometric distortions, field of view, and chromatic aberration are achieved without the need for a priori knowledge of the display system's intrinsic parameters. This method operates by capturing and storing the pixel space locations of a series of real world control points. These control points are then used as ground truth references by which virtual space transformations can be automatically generated for a display system.", "abstracts": [ { "abstractType": "Regular", "content": "In this document we present a method for calibrating head-mounted displays and other display surfaces using an automated, low-cost camera system. A unique aspect of this method is that the calibration of geometric distortions, field of view, and chromatic aberration are achieved without the need for a priori knowledge of the display system's intrinsic parameters. This method operates by capturing and storing the pixel space locations of a series of real world control points. These control points are then used as ground truth references by which virtual space transformations can be automatically generated for a display system.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this document we present a method for calibrating head-mounted displays and other display surfaces using an automated, low-cost camera system. A unique aspect of this method is that the calibration of geometric distortions, field of view, and chromatic aberration are achieved without the need for a priori knowledge of the display system's intrinsic parameters. This method operates by capturing and storing the pixel space locations of a series of real world control points. These control points are then used as ground truth references by which virtual space transformations can be automatically generated for a display system.", "fno": "06802063", "keywords": [ "Cameras", "Calibration", "Aerospace Electronics", "Image Color Analysis", "Virtual Environments", "Optical Surface Waves", "Educational Institutions" ], "authors": [ { "affiliation": "University of Southern California", "fullName": "J. Adam Jones", "givenName": "J. Adam", "surname": "Jones", "__typename": "ArticleAuthorType" }, { "affiliation": "Clemson University", "fullName": "Lauren Cairco Dukes", "givenName": "Lauren Cairco", "surname": "Dukes", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Southern California", "fullName": "Mark Bolas", "givenName": "Mark", "surname": "Bolas", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-03-01T00:00:00", "pubType": "proceedings", "pages": "85-86", "year": "2014", "issn": null, "isbn": "978-1-4799-2871-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06802062", "articleId": "12OmNvjgWS5", "__typename": "AdjacentArticleType" }, "next": { "fno": "06802064", "articleId": "12OmNCcbDXX", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ismar/2015/7660/0/7660a043", "title": "Simultaneous Direct and Augmented View Distortion Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2015/7660a043/12OmNC1oT64", "parentPublication": { "id": "proceedings/ismar/2015/7660/0", "title": "2015 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2015/1727/0/07223443", "title": "Non-obscuring binocular eye tracking for wide field-of-view head-mounted-displays", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223443/12OmNqzu6MP", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1993/1363/0/00380772", "title": "Calibration of head-mounted displays for augmented reality applications", "doi": null, "abstractUrl": "/proceedings-article/vrais/1993/00380772/12OmNwwuDRm", "parentPublication": { "id": "proceedings/vrais/1993/1363/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2013/4795/0/06549395", "title": "Flexible and general redirected walking for head-mounted displays", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549395/12OmNxFJXN3", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iwar/1999/0359/0/03590075", "title": "A Method for Calibrating See-Through Head-Mounted Displays for AR", "doi": null, "abstractUrl": "/proceedings-article/iwar/1999/03590075/12OmNxTVU20", "parentPublication": { "id": "proceedings/iwar/1999/0359/0", "title": "Augmented Reality, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446429", "title": "Impact of Alignment Point Distance Distribution on SPAAM Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446429/13bd1gCd7Sz", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07012105", "title": "Corneal-Imaging Calibration for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/04/07012105/13rRUxE04tC", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2022/5325/0/532500a131", "title": "Evaluation of Text Selection Techniques in Virtual Reality Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2022/532500a131/1JrRdnGe43C", "parentPublication": { "id": "proceedings/ismar/2022/5325/0", "title": "2022 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2023/05/10050791", "title": "Add-on Occlusion: Turning Off-the-Shelf Optical See-through Head-mounted Displays Occlusion-capable", "doi": null, "abstractUrl": "/journal/tg/2023/05/10050791/1L039oS5wDm", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2021/0158/0/015800a001", "title": "Perception-Driven Hybrid Foveated Depth of Field Rendering for Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2021/015800a001/1yeCURkWXpS", "parentPublication": { "id": "proceedings/ismar/2021/0158/0", "title": "2021 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1gyshXRzHpK", "title": "2019 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "acronym": "ismar-adjunct", "groupId": "1810084", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1gysj1o4L16", "doi": "10.1109/ISMAR-Adjunct.2019.00-34", "title": "OSTNet: Calibration Method for Optical See-Through Head-Mounted Displays via Non-Parametric Distortion Map Generation", "normalizedTitle": "OSTNet: Calibration Method for Optical See-Through Head-Mounted Displays via Non-Parametric Distortion Map Generation", "abstract": "We propose a spatial calibration method for Optical See-Through Head-Mounted Displays (OST-HMDs) having complex optical distortion such as wide field-of-view (FoV) designs. Viewpoint-dependent non-linear optical distortion makes existing spatial calibration methods either impossible to handle or difficult to compensate without intensive computation. To overcome this issue, we propose OSTNet, a non-parametric data-driven calibration method that creates a generative 2D distortion model for a given six-degree-of-freedom viewpoint pose.", "abstracts": [ { "abstractType": "Regular", "content": "We propose a spatial calibration method for Optical See-Through Head-Mounted Displays (OST-HMDs) having complex optical distortion such as wide field-of-view (FoV) designs. Viewpoint-dependent non-linear optical distortion makes existing spatial calibration methods either impossible to handle or difficult to compensate without intensive computation. To overcome this issue, we propose OSTNet, a non-parametric data-driven calibration method that creates a generative 2D distortion model for a given six-degree-of-freedom viewpoint pose.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose a spatial calibration method for Optical See-Through Head-Mounted Displays (OST-HMDs) having complex optical distortion such as wide field-of-view (FoV) designs. Viewpoint-dependent non-linear optical distortion makes existing spatial calibration methods either impossible to handle or difficult to compensate without intensive computation. To overcome this issue, we propose OSTNet, a non-parametric data-driven calibration method that creates a generative 2D distortion model for a given six-degree-of-freedom viewpoint pose.", "fno": "476500a259", "keywords": [ "Calibration", "Helmet Mounted Displays", "Optical Distortion", "Nonparametric Distortion Map Generation", "Spatial Calibration Method", "Complex Optical Distortion", "Field Of View Designs", "Viewpoint Dependent Nonlinear Optical Distortion", "Nonparametric Data Driven Calibration Method", "Generative 2 D Distortion Model", "OST Net", "Optical Distortion", "Calibration", "Cameras", "Nonlinear Distortion", "Two Dimensional Displays", "Decoding", "Optical See Through Head Mounted Display", "Calibration", "Variational Autoencoder" ], "authors": [ { "affiliation": "Tokyo Institute of Technology", "fullName": "Kiyosato Someya", "givenName": "Kiyosato", "surname": "Someya", "__typename": "ArticleAuthorType" }, { "affiliation": "Keio University", "fullName": "Yuichi Hiroi", "givenName": "Yuichi", "surname": "Hiroi", "__typename": "ArticleAuthorType" }, { "affiliation": "Kyoto University", "fullName": "Makoto Yamada", "givenName": "Makoto", "surname": "Yamada", "__typename": "ArticleAuthorType" }, { "affiliation": "Tokyo Institute of Technology", "fullName": "Yuta Itoh", "givenName": "Yuta", "surname": "Itoh", "__typename": "ArticleAuthorType" } ], "idPrefix": "ismar-adjunct", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-10-01T00:00:00", "pubType": "proceedings", "pages": "259-260", "year": "2019", "issn": null, "isbn": "978-1-7281-4765-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "476500a257", "articleId": "1gysnSDS8Wk", "__typename": "AdjacentArticleType" }, "next": { "fno": "476500a261", "articleId": "1gysikN6QOQ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2016/0836/0/07504693", "title": "A calibration method for optical see-through head-mounted displays with a depth camera", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504693/12OmNAnMuMd", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948513", "title": "Google glass, The META and Co. How to calibrate optical see-through head mounted displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948513/12OmNB8TUim", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2015/7660/0/7660a043", "title": "Simultaneous Direct and Augmented View Distortion Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2015/7660a043/12OmNC1oT64", "parentPublication": { "id": "proceedings/ismar/2015/7660/0", "title": "2015 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2013/4795/0/06549353", "title": "A robust camera-based method for optical distortion calibration of head-mounted displays", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549353/12OmNwvVrHy", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2014/6184/0/06948424", "title": "Performance and sensitivity analysis of INDICA: INteraction-Free DIsplay CAlibration for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar/2014/06948424/12OmNyYm2oO", "parentPublication": { "id": "proceedings/ismar/2014/6184/0", "title": "2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2015/1727/0/07223385", "title": "Continuous automatic calibration for optical see-through displays", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223385/12OmNynJMQZ", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/09/08052554", "title": "A Survey of Calibration Methods for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2018/09/08052554/13rRUILtJqY", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/11/07165643", "title": "Semi-Parametric Color Reproduction Method for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/11/07165643/13rRUILtJzB", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07064856", "title": "Light-Field Correction for Spatial Calibration of Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2015/04/07064856/13rRUwjGoG5", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/11/07523375", "title": "Gaussian Light Field: Estimation of Viewpoint-Dependent Blur for Optical See-Through Head-Mounted Displays", "doi": null, "abstractUrl": "/journal/tg/2016/11/07523375/13rRUxYINfi", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNrIJqww", "title": "2013 26th Conference on Graphics, Patterns and Images Tutorials (SIBGRAPI-T)", "acronym": "sibgrapi-t", "groupId": "1003086", "volume": "0", "displayVolume": "0", "year": "2013", "__typename": "ProceedingType" }, "article": { "id": "12OmNAmVH6E", "doi": "10.1109/SIBGRAPI-T.2013.9", "title": "Visual Immersion Issues in Virtual Reality: A Survey", "normalizedTitle": "Visual Immersion Issues in Virtual Reality: A Survey", "abstract": "Thanks to immersion and interaction, Virtual Reality (VR) offers a shift of paradigm compared to traditional computer graphics or simulation software. Most VR applications include a visual rendering part. However, efficient and relevant visual interfacing of a human user (who can be a designer, the final user of a future product, a trainee or the subject of an experiment) raises issues about visual interfaces and depth perception in computer generated images. In this survey we review the issues raised by the visual part of VR applications. We particularly focus on the commonly used stereoscopic vision by studying its constraints and show how an efficient stereoscopic application should be designed.", "abstracts": [ { "abstractType": "Regular", "content": "Thanks to immersion and interaction, Virtual Reality (VR) offers a shift of paradigm compared to traditional computer graphics or simulation software. Most VR applications include a visual rendering part. However, efficient and relevant visual interfacing of a human user (who can be a designer, the final user of a future product, a trainee or the subject of an experiment) raises issues about visual interfaces and depth perception in computer generated images. In this survey we review the issues raised by the visual part of VR applications. We particularly focus on the commonly used stereoscopic vision by studying its constraints and show how an efficient stereoscopic application should be designed.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Thanks to immersion and interaction, Virtual Reality (VR) offers a shift of paradigm compared to traditional computer graphics or simulation software. Most VR applications include a visual rendering part. However, efficient and relevant visual interfacing of a human user (who can be a designer, the final user of a future product, a trainee or the subject of an experiment) raises issues about visual interfaces and depth perception in computer generated images. In this survey we review the issues raised by the visual part of VR applications. We particularly focus on the commonly used stereoscopic vision by studying its constraints and show how an efficient stereoscopic application should be designed.", "fno": "4176a006", "keywords": [ "Visualization", "Stereo Image Processing", "Virtual Reality", "Head", "Observers", "Rendering Computer Graphics", "Convergence", "Visual Fatigue", "Virtual Reality", "Immersion", "Interaction", "Depth Perception", "Stereoscopic Viewing" ], "authors": [ { "affiliation": null, "fullName": "Guillaume Moreau", "givenName": "Guillaume", "surname": "Moreau", "__typename": "ArticleAuthorType" } ], "idPrefix": "sibgrapi-t", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2013-08-01T00:00:00", "pubType": "proceedings", "pages": "6-14", "year": "2013", "issn": null, "isbn": "978-1-4799-4176-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4176a001", "articleId": "12OmNqGA54v", "__typename": "AdjacentArticleType" }, "next": { "fno": "4176a015", "articleId": "12OmNzBOi3K", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2018/3365/0/08446222", "title": "A Method of View-Dependent Stereoscopic Projection on Curved Screen", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446222/13bd1gCd7Sx", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2015/05/mcg2015050034", "title": "Reducing Visual Discomfort with HMDs Using Dynamic Depth of Field", "doi": null, "abstractUrl": "/magazine/cg/2015/05/mcg2015050034/13rRUEgarvh", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040597", "title": "Effects of Immersion on Visual Analysis of Volume Data", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040597/13rRUxASuve", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/04/08269807", "title": "Saliency in VR: How Do People Explore Virtual Environments?", "doi": null, "abstractUrl": "/journal/tg/2018/04/08269807/13rRUxDqS8o", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/11/09872532", "title": "FoV-NeRF: Foveated Neural Radiance Fields for Virtual Reality", "doi": null, "abstractUrl": "/journal/tg/2022/11/09872532/1GhRVOIAS4g", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08797876", "title": "Live Stereoscopic 3D Image with Constant Capture Direction of 360&#x00B0; Cameras for High-Quality Visual Telepresence", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08797876/1cJ0HMTqjOU", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798112", "title": "The Effects of Stereopsis and Immersion on Bimanual Assembly Tasks in a Virtual Reality System", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798112/1cJ0WXAmYKY", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798283", "title": "Hybrid Mono-Stereo Rendering in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798283/1cJ11Kv4Dn2", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vs-games/2019/4540/0/08864578", "title": "Mono-Stereoscopic Camera in a Virtual Reality Environment: Case Study in Cybersickness", "doi": null, "abstractUrl": "/proceedings-article/vs-games/2019/08864578/1e5Zs94AhSE", "parentPublication": { "id": "proceedings/vs-games/2019/4540/0", "title": "2019 11th International Conference on Virtual Worlds and Games for Serious Applications (VS-Games)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089475", "title": "Recurrent Enhancement of Visual Comfort for Casual Stereoscopic Photography", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089475/1jIx9ixQvzq", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyPQ4vy", "title": "2009 International Conference on Network-Based Information Systems", "acronym": "nbis", "groupId": "1002969", "volume": "0", "displayVolume": "0", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNB06l11", "doi": "10.1109/NBiS.2009.89", "title": "Tele-Immersion Environment for Video Avatar Based CVE", "normalizedTitle": "Tele-Immersion Environment for Video Avatar Based CVE", "abstract": "In this paper, we propose a middleware system to realize collaborative virtual environment (CVE) considering tele-immersion environment based on personal computer. Using this middleware system, more immersive and presence feeling is attained and collaborative work can be effectively improved compared with using immersive display like CAVE. We built a prototype system of the proposed method and carried out functional and performance evaluations. As a result, we could verify that our proposed system is very cost effective and provided basic CVE functions on PC environment.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, we propose a middleware system to realize collaborative virtual environment (CVE) considering tele-immersion environment based on personal computer. Using this middleware system, more immersive and presence feeling is attained and collaborative work can be effectively improved compared with using immersive display like CAVE. We built a prototype system of the proposed method and carried out functional and performance evaluations. As a result, we could verify that our proposed system is very cost effective and provided basic CVE functions on PC environment.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, we propose a middleware system to realize collaborative virtual environment (CVE) considering tele-immersion environment based on personal computer. Using this middleware system, more immersive and presence feeling is attained and collaborative work can be effectively improved compared with using immersive display like CAVE. We built a prototype system of the proposed method and carried out functional and performance evaluations. As a result, we could verify that our proposed system is very cost effective and provided basic CVE functions on PC environment.", "fno": "3767a608", "keywords": [ "Tele Immersion", "Virtual Reality" ], "authors": [ { "affiliation": null, "fullName": "Tomoyuki Ishida", "givenName": "Tomoyuki", "surname": "Ishida", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Keisuke Yatsu", "givenName": "Keisuke", "surname": "Yatsu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yoshitaka Shibata", "givenName": "Yoshitaka", "surname": "Shibata", "__typename": "ArticleAuthorType" } ], "idPrefix": "nbis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-08-01T00:00:00", "pubType": "proceedings", "pages": "608-611", "year": "2009", "issn": null, "isbn": "978-0-7695-3767-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3767a603", "articleId": "12OmNyRxFrK", "__typename": "AdjacentArticleType" }, "next": { "fno": "3767a618", "articleId": "12OmNzYNNlt", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2011/0039/0/05759470", "title": "VEGI: Virtual Environment GUI Immersion system", "doi": null, "abstractUrl": "/proceedings-article/vr/2011/05759470/12OmNBKEyn1", "parentPublication": { "id": "proceedings/vr/2011/0039/0", "title": "2011 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hpdc/1999/0287/0/02870008", "title": "A Methodology for Supporting Collaborative Exploratory Analysis of Massive Data Sets in Tele-Immersive Environments", "doi": null, "abstractUrl": "/proceedings-article/hpdc/1999/02870008/12OmNCbkQB4", "parentPublication": { "id": "proceedings/hpdc/1999/0287/0", "title": "High-Performance Distributed Computing, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/waina/2009/3639/0/3639a554", "title": "An Experiment on Tele-immersive Communication with Tiled Displays Wall over JGN2plus Network", "doi": null, "abstractUrl": "/proceedings-article/waina/2009/3639a554/12OmNqJ8tuG", "parentPublication": { "id": "proceedings/waina/2009/3639/0", "title": "2009 International Conference on Advanced Information Networking and Applications Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ainaw/2008/3096/0/3096b023", "title": "Proposal of Multimodal Communications System Using Tele-Immersion Technology", "doi": null, "abstractUrl": "/proceedings-article/ainaw/2008/3096b023/12OmNvHGrAg", "parentPublication": { "id": "proceedings/ainaw/2008/3096/0", "title": "2008 22nd International Conference on Advanced Information Networking and Applications (AINA 2008)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icat/2006/2754/0/27540131", "title": "MOPCVE: A New CVE Prototype Running on Mobile Phone", "doi": null, "abstractUrl": "/proceedings-article/icat/2006/27540131/12OmNxG1yWl", "parentPublication": { "id": "proceedings/icat/2006/2754/0", "title": "16th International Conference on Artificial Reality and Telexistence--Workshops (ICAT'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2008/3454/0/3454a477", "title": "Implementing a Distributed Tele-immersive System", "doi": null, "abstractUrl": "/proceedings-article/ism/2008/3454a477/12OmNxIzWOu", "parentPublication": { "id": "proceedings/ism/2008/3454/0", "title": "2008 Tenth IEEE International Symposium on Multimedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nbis/2010/4167/0/4167a424", "title": "Asynchronous Collaborative Support System Introducing Access Control in CVE", "doi": null, "abstractUrl": "/proceedings-article/nbis/2010/4167a424/12OmNyRPgUf", "parentPublication": { "id": "proceedings/nbis/2010/4167/0", "title": "2010 13th International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/waina/2009/3639/0/3639a560", "title": "A Middleware System to Realize Virtual Reality on Tele-immersion Environment", "doi": null, "abstractUrl": "/proceedings-article/waina/2009/3639a560/12OmNybfqXn", "parentPublication": { "id": "proceedings/waina/2009/3639/0", "title": "2009 International Conference on Advanced Information Networking and Applications Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444767", "title": "VirtualizeMe: Real-time avatar creation for Tele-Immersion environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444767/12OmNzahcfu", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nbis/2010/4167/0/4167a408", "title": "Proposal of Tele-immersion System by the Fusion of Virtual Space and Real Space", "doi": null, "abstractUrl": "/proceedings-article/nbis/2010/4167a408/12OmNzn390Q", "parentPublication": { "id": "proceedings/nbis/2010/4167/0", "title": "2010 13th International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzw8jgZ", "title": "2011 IEEE Virtual Reality (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNBKEyn1", "doi": "10.1109/VR.2011.5759470", "title": "VEGI: Virtual Environment GUI Immersion system", "normalizedTitle": "VEGI: Virtual Environment GUI Immersion system", "abstract": "Virtual Reality (VR) immersive environments are becoming more popular and of less cost, hence, VR labs are becoming a main part in any research that depends on visualization. This introduced the need to port many 3D desktop visualization applications to VR. Porting application GUIs can be a problem since original GUIs are 2D by nature and using them directly can obscure a large area of 3D viewport and spoil the immersive experience. On the other hand, rewriting a 3D GUI can be a time consuming and tedious task. In this work, we introduce a technique to embed 2D GUIs into 3D Virtual Environments (VE). Our approach uses existing 2D GUIs that can be immersed into the VE allowing rapid GUI development for VR applications. It can also be used for porting 3D desktop applications without rewriting the GUI code. Further, it enables embedding many window-based desktop applications into the VE, creating rich VEs where users can work with multiple applications simultaneously.", "abstracts": [ { "abstractType": "Regular", "content": "Virtual Reality (VR) immersive environments are becoming more popular and of less cost, hence, VR labs are becoming a main part in any research that depends on visualization. This introduced the need to port many 3D desktop visualization applications to VR. Porting application GUIs can be a problem since original GUIs are 2D by nature and using them directly can obscure a large area of 3D viewport and spoil the immersive experience. On the other hand, rewriting a 3D GUI can be a time consuming and tedious task. In this work, we introduce a technique to embed 2D GUIs into 3D Virtual Environments (VE). Our approach uses existing 2D GUIs that can be immersed into the VE allowing rapid GUI development for VR applications. It can also be used for porting 3D desktop applications without rewriting the GUI code. Further, it enables embedding many window-based desktop applications into the VE, creating rich VEs where users can work with multiple applications simultaneously.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Virtual Reality (VR) immersive environments are becoming more popular and of less cost, hence, VR labs are becoming a main part in any research that depends on visualization. This introduced the need to port many 3D desktop visualization applications to VR. Porting application GUIs can be a problem since original GUIs are 2D by nature and using them directly can obscure a large area of 3D viewport and spoil the immersive experience. On the other hand, rewriting a 3D GUI can be a time consuming and tedious task. In this work, we introduce a technique to embed 2D GUIs into 3D Virtual Environments (VE). Our approach uses existing 2D GUIs that can be immersed into the VE allowing rapid GUI development for VR applications. It can also be used for porting 3D desktop applications without rewriting the GUI code. Further, it enables embedding many window-based desktop applications into the VE, creating rich VEs where users can work with multiple applications simultaneously.", "fno": "05759470", "keywords": [ "Window Based Desktop Application", "Virtual Environment GUI Immersion System", "VEGI", "Virtual Reality Immersive Environment", "VR Lab", "3 D Desktop Visualization Application", "3 D GUI", "3 D Virtual Environment" ], "authors": [ { "affiliation": "Bibliotheca Alexandrina, Alexandria, Egypt", "fullName": "M Elfarargy", "givenName": "M", "surname": "Elfarargy", "__typename": "ArticleAuthorType" }, { "affiliation": "Bibliotheca Alexandrina, Alexandria, Egypt", "fullName": "M Nagi", "givenName": "M", "surname": "Nagi", "__typename": "ArticleAuthorType" }, { "affiliation": "Bibliotheca Alexandrina, Alexandria, Egypt", "fullName": "N Adly", "givenName": "N", "surname": "Adly", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-03-01T00:00:00", "pubType": "proceedings", "pages": "205-206", "year": "2011", "issn": null, "isbn": "978-1-4577-0039-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05759469", "articleId": "12OmNAlvHIu", "__typename": "AdjacentArticleType" }, "next": { "fno": "05759471", "articleId": "12OmNqGiu1d", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icstw/2010/4050/0/4050a240", "title": "Using Methods & Measures from Network Analysis for GUI Testing", "doi": null, "abstractUrl": "/proceedings-article/icstw/2010/4050a240/12OmNBr4exQ", "parentPublication": { "id": "proceedings/icstw/2010/4050/0", "title": "Software Testing Verification and Validation Workshop, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iat/2005/2416/0/24160165", "title": "Actionable Knowledge Model for GUI Regression Testing", "doi": null, "abstractUrl": "/proceedings-article/iat/2005/24160165/12OmNC4O4Bi", "parentPublication": { "id": "proceedings/iat/2005/2416/0", "title": "Proceedings. The 2005 IEEE/WIC/ACM International Conference on Intelligent Agent Technology", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/taic-part/2008/3383/0/3383a023", "title": "Alternating GUI Test Generation and Execution", "doi": null, "abstractUrl": "/proceedings-article/taic-part/2008/3383a023/12OmNx19k2u", "parentPublication": { "id": "proceedings/taic-part/2008/3383/0", "title": "Practice And Research Techniques, Testing: Academic & Industrial Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/issre/2001/1306/0/13060054", "title": "User-Based Testing of GUI Sequences and Their Interactions", "doi": null, "abstractUrl": "/proceedings-article/issre/2001/13060054/12OmNxFsmCn", "parentPublication": { "id": "proceedings/issre/2001/1306/0", "title": "Proceedings 12th International Symposium on Software Reliability Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icsm/2003/1905/0/19050398", "title": "Firewall Regression Testing of GUI Sequences and their Interactions", "doi": null, "abstractUrl": "/proceedings-article/icsm/2003/19050398/12OmNyKJifD", "parentPublication": { "id": "proceedings/icsm/2003/1905/0", "title": "International Conference on Software Maintenance, 2003. ICSM 2003. Proceedings.", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/taic-part/2008/3383/0/3383a099", "title": "GUI Testing Made Easy", "doi": null, "abstractUrl": "/proceedings-article/taic-part/2008/3383a099/12OmNyPQ4It", "parentPublication": { "id": "proceedings/taic-part/2008/3383/0", "title": "Practice And Research Techniques, Testing: Academic & Industrial Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ase/2012/1204/0/2351739", "title": "GUITest: a Java library for fully automated GUI robustness testing", "doi": null, "abstractUrl": "/proceedings-article/ase/2012/2351739/12OmNzG4gt3", "parentPublication": { "id": "proceedings/ase/2012/1204/0", "title": "2012 Proceedings of the 27th IEEE/ACM International Conference on Automated Software Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/compsac/2009/3726/1/3726a632", "title": "An Ontology-Based Approach for GUI Testing", "doi": null, "abstractUrl": "/proceedings-article/compsac/2009/3726a632/12OmNzsJ7yt", "parentPublication": { "id": "proceedings/compsac/2009/3726/1", "title": "2009 33rd Annual IEEE International Computer Software and Applications Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ts/2001/02/e0144", "title": "Hierarchical GUI Test Case Generation Using Automated Planning", "doi": null, "abstractUrl": "/journal/ts/2001/02/e0144/13rRUxNEqRM", "parentPublication": { "id": "trans/ts", "title": "IEEE Transactions on Software Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/so/2007/03/s3051", "title": "Test-Driven GUI Development with TestNG and Abbot", "doi": null, "abstractUrl": "/magazine/so/2007/03/s3051/13rRUy0HYPi", "parentPublication": { "id": "mags/so", "title": "IEEE Software", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNySXF3b", "title": "2008 22nd International Conference on Advanced Information Networking and Applications (AINA 2008)", "acronym": "ainaw", "groupId": "1001766", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNvHGrAg", "doi": "10.1109/WAINA.2008.153", "title": "Proposal of Multimodal Communications System Using Tele-Immersion Technology", "normalizedTitle": "Proposal of Multimodal Communications System Using Tele-Immersion Technology", "abstract": "In this paper, it proposes the Multimodal Communications System that uses three information on Verbal Information, Non-verbal Information, and Sensitivity Information. In this research, universities and research institutes that have Immersive Displays and Tiled Displays are interconnected over the ultrahigh-speed network, and cooperative Tele-Immersion environments are constructed. A new human community is created by using Tele-Immersion Technology of this research. The facial expression and the gesture are the critical factors in person's communications, and enough communications might not be approved in present video phone and teleconference system. The multimodal communications system achieves new communications in interconnected large screen system by using the video communication technology and the sensibility information processing technology.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, it proposes the Multimodal Communications System that uses three information on Verbal Information, Non-verbal Information, and Sensitivity Information. In this research, universities and research institutes that have Immersive Displays and Tiled Displays are interconnected over the ultrahigh-speed network, and cooperative Tele-Immersion environments are constructed. A new human community is created by using Tele-Immersion Technology of this research. The facial expression and the gesture are the critical factors in person's communications, and enough communications might not be approved in present video phone and teleconference system. The multimodal communications system achieves new communications in interconnected large screen system by using the video communication technology and the sensibility information processing technology.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, it proposes the Multimodal Communications System that uses three information on Verbal Information, Non-verbal Information, and Sensitivity Information. In this research, universities and research institutes that have Immersive Displays and Tiled Displays are interconnected over the ultrahigh-speed network, and cooperative Tele-Immersion environments are constructed. A new human community is created by using Tele-Immersion Technology of this research. The facial expression and the gesture are the critical factors in person's communications, and enough communications might not be approved in present video phone and teleconference system. The multimodal communications system achieves new communications in interconnected large screen system by using the video communication technology and the sensibility information processing technology.", "fno": "3096b023", "keywords": [ "Multimodal Communications", "Tele Immersion", "Virtual Reality" ], "authors": [ { "affiliation": null, "fullName": "Tomoyuki Ishida", "givenName": "Tomoyuki", "surname": "Ishida", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Akihiro Miyakawa", "givenName": "Akihiro", "surname": "Miyakawa", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yoshitaka Shibata", "givenName": "Yoshitaka", "surname": "Shibata", "__typename": "ArticleAuthorType" } ], "idPrefix": "ainaw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-03-01T00:00:00", "pubType": "proceedings", "pages": "1023-1028", "year": "2008", "issn": null, "isbn": "978-0-7695-3096-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3096b017", "articleId": "12OmNzYNNjD", "__typename": "AdjacentArticleType" }, "next": { "fno": "3096b029", "articleId": "12OmNvlPkAt", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/nbis/2009/3767/0/3767a608", "title": "Tele-Immersion Environment for Video Avatar Based CVE", "doi": null, "abstractUrl": "/proceedings-article/nbis/2009/3767a608/12OmNB06l11", "parentPublication": { "id": "proceedings/nbis/2009/3767/0", "title": "2009 International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/eee/2005/2274/0/22740646", "title": "A Service-Oriented Architecture for Tele-Immersion", "doi": null, "abstractUrl": "/proceedings-article/eee/2005/22740646/12OmNC4wtME", "parentPublication": { "id": "proceedings/eee/2005/2274/0", "title": "Proceedings. The 2005 IEEE International Conference on e-Technology, e-Commerce and e-Service", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2008/3454/0/3454a477", "title": "Implementing a Distributed Tele-immersive System", "doi": null, "abstractUrl": "/proceedings-article/ism/2008/3454a477/12OmNxIzWOu", "parentPublication": { "id": "proceedings/ism/2008/3454/0", "title": "2008 Tenth IEEE International Symposium on Multimedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdcsw/2008/3173/0/3173a072", "title": "Collaborative and Multimodal Communications System Using Immersive Virtual Reality Environment over Ultrahigh-Speed Network", "doi": null, "abstractUrl": "/proceedings-article/icdcsw/2008/3173a072/12OmNxw5B5z", "parentPublication": { "id": "proceedings/icdcsw/2008/3173/0", "title": "2008 The 28th International Conference on Distributed Computing Systems Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3pgcic/2010/4237/0/4237a421", "title": "An Experiment on Multipoint Tele-immersive Communication with Tiled Display Wall", "doi": null, "abstractUrl": "/proceedings-article/3pgcic/2010/4237a421/12OmNy2ah07", "parentPublication": { "id": "proceedings/3pgcic/2010/4237/0", "title": "P2P, Parallel, Grid, Cloud, and Internet Computing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2012/4875/0/4875a378", "title": "TEEVE-Remote: A Novel User-Interaction Solution for 3D Tele-immersive System", "doi": null, "abstractUrl": "/proceedings-article/ism/2012/4875a378/12OmNylbosj", "parentPublication": { "id": "proceedings/ism/2012/4875/0", "title": "2012 IEEE International Symposium on Multimedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hicss/2000/0493/5/04935021", "title": "Virtual Reality and Tele-Immersion in Health Care Introduction to Minitrack", "doi": null, "abstractUrl": "/proceedings-article/hicss/2000/04935021/12OmNyyO8OC", "parentPublication": { "id": "proceedings/hicss/2000/0493/5", "title": "Proceedings of the 33rd Annual Hawaii International Conference on System Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444767", "title": "VirtualizeMe: Real-time avatar creation for Tele-Immersion environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444767/12OmNzahcfu", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nbis/2010/4167/0/4167a408", "title": "Proposal of Tele-immersion System by the Fusion of Virtual Space and Real Space", "doi": null, "abstractUrl": "/proceedings-article/nbis/2010/4167a408/12OmNzn390Q", "parentPublication": { "id": "proceedings/nbis/2010/4167/0", "title": "2010 13th International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/10/08093688", "title": "Tele-Immersion Concepts", "doi": null, "abstractUrl": "/journal/tg/2018/10/08093688/13rRUxcsYLZ", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1cI6akLvAuQ", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1cJ16iGCIxy", "doi": "10.1109/VR.2019.8798021", "title": "The Effects of Immersion on Harm-inducing Factors in Virtual Slot Machines", "normalizedTitle": "The Effects of Immersion on Harm-inducing Factors in Virtual Slot Machines", "abstract": "Slot machines are one of the most played games by pathological gamblers. New technologies, e.g. immersive Virtual Reality (VR), offer more possibilities to exploit erroneous beliefs in the context of gambling. However, the risk potential of VR-based gambling has not been researched, yet. A higher immersion might increase harmful aspects, thus making VR realizations more dangerous. Measuring harm-inducing factors reveals the risk potential of virtual gambling. In a user study, we analyze a slot machine realized as a desktop 3D and as an immersive VR version. Both versions are compared in respect to effects on dissociation, urge to gamble, dark flow, and illusion of control. Our study shows significantly higher values of dissociation, dark flow, and urge to gamble in the VR version. Presence significantly correlates with all measured harm-inducing factors. We demonstrate that VR-based gambling has a higher risk potential. This creates the importance of regulating VR-based gambling.", "abstracts": [ { "abstractType": "Regular", "content": "Slot machines are one of the most played games by pathological gamblers. New technologies, e.g. immersive Virtual Reality (VR), offer more possibilities to exploit erroneous beliefs in the context of gambling. However, the risk potential of VR-based gambling has not been researched, yet. A higher immersion might increase harmful aspects, thus making VR realizations more dangerous. Measuring harm-inducing factors reveals the risk potential of virtual gambling. In a user study, we analyze a slot machine realized as a desktop 3D and as an immersive VR version. Both versions are compared in respect to effects on dissociation, urge to gamble, dark flow, and illusion of control. Our study shows significantly higher values of dissociation, dark flow, and urge to gamble in the VR version. Presence significantly correlates with all measured harm-inducing factors. We demonstrate that VR-based gambling has a higher risk potential. This creates the importance of regulating VR-based gambling.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Slot machines are one of the most played games by pathological gamblers. New technologies, e.g. immersive Virtual Reality (VR), offer more possibilities to exploit erroneous beliefs in the context of gambling. However, the risk potential of VR-based gambling has not been researched, yet. A higher immersion might increase harmful aspects, thus making VR realizations more dangerous. Measuring harm-inducing factors reveals the risk potential of virtual gambling. In a user study, we analyze a slot machine realized as a desktop 3D and as an immersive VR version. Both versions are compared in respect to effects on dissociation, urge to gamble, dark flow, and illusion of control. Our study shows significantly higher values of dissociation, dark flow, and urge to gamble in the VR version. Presence significantly correlates with all measured harm-inducing factors. We demonstrate that VR-based gambling has a higher risk potential. This creates the importance of regulating VR-based gambling.", "fno": "08798021", "keywords": [ "Human Factors", "Humanities", "Virtual Reality", "Virtual Slot Machines", "VR Based Gambling", "VR Realizations", "Virtual Gambling", "Immersive VR Version", "Immersive Virtual Reality", "Games", "Pathology", "Virtual Reality", "Three Dimensional Displays", "Indexes", "Visualization", "Current Measurement", "Human Centered Computing X 2014 Human Computer Interaction HCI X 2014 HCI Design And Evaluation Methods", "Human Centered Computing X 2014 Human Computer Interaction HCI X 2014 Empirical Studies In HCI", "Human Centered Computing X 2014 Interaction Paradigms X 2014 Virtual Reality" ], "authors": [ { "affiliation": "University of Würzburg", "fullName": "David Heidrich", "givenName": "David", "surname": "Heidrich", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Würzburg", "fullName": "Sebastian Oberdörfer", "givenName": "Sebastian", "surname": "Oberdörfer", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Würzburg", "fullName": "Marc Erich Latoschik", "givenName": "Marc Erich", "surname": "Latoschik", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-03-01T00:00:00", "pubType": "proceedings", "pages": "793-801", "year": "2019", "issn": null, "isbn": "978-1-7281-1377-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08798208", "articleId": "1cJ0JhaLbzi", "__typename": "AdjacentArticleType" }, "next": { "fno": "08797977", "articleId": "1cJ0OGvQaSQ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ismar-adjunct/2017/6327/0/6327a310", "title": "Improving the Immersion in Virtual Reality with Real-Time Avatar and Haptic Feedback in a Cricket Simulation", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2017/6327a310/12OmNyGbIfg", "parentPublication": { "id": "proceedings/ismar-adjunct/2017/6327/0", "title": "2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446221", "title": "Simulator Sick but Still Immersed: A Comparison of Head-Object Collision Handling and Their Impact on Fun, Immersion, and Simulator Sickness", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446221/13bd1eSlytz", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08797899", "title": "A VR Interactive Story Using POV and Flashback for Empathy", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08797899/1cJ12bET8XK", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089482", "title": "Think Twice: The Influence of Immersion on Decision Making during Gambling in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089482/1jIxeHgBEkg", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1tnWwqMuCzu", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "acronym": "vrw", "groupId": "1836626", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1tnY3EP4yM8", "doi": "10.1109/VRW52623.2021.00102", "title": "Effects of Immersion and Visual Angle on Brand Placement Effectiveness", "normalizedTitle": "Effects of Immersion and Visual Angle on Brand Placement Effectiveness", "abstract": "Typical inherent properties of immersive Virtual Reality (VR) such as felt presence might have an impact on how well brand placements are remembered. In this study, we exposed participants to brand placements in four conditions of varying degrees of immersion and visual angle on the stimulus. Placements appeared either as poster or as puzzle. We measured the recall and recognition of these placements. Our study revealed that neither immersion nor the visual angle had a significant impact on memory for brand placements.", "abstracts": [ { "abstractType": "Regular", "content": "Typical inherent properties of immersive Virtual Reality (VR) such as felt presence might have an impact on how well brand placements are remembered. In this study, we exposed participants to brand placements in four conditions of varying degrees of immersion and visual angle on the stimulus. Placements appeared either as poster or as puzzle. We measured the recall and recognition of these placements. Our study revealed that neither immersion nor the visual angle had a significant impact on memory for brand placements.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Typical inherent properties of immersive Virtual Reality (VR) such as felt presence might have an impact on how well brand placements are remembered. In this study, we exposed participants to brand placements in four conditions of varying degrees of immersion and visual angle on the stimulus. Placements appeared either as poster or as puzzle. We measured the recall and recognition of these placements. Our study revealed that neither immersion nor the visual angle had a significant impact on memory for brand placements.", "fno": "405700a440", "keywords": [ "Human Computer Interaction", "Marketing", "Virtual Reality", "Visual Angle", "Brand Placement Effectiveness", "Typical Inherent Properties", "Immersive Virtual Reality", "Felt Presence", "Brand Placements", "VR", "Visualization", "Three Dimensional Displays", "Atmospheric Measurements", "Conferences", "Design Methodology", "Virtual Reality", "User Interfaces", "Human Centered Computing", "Human Computer Interaction HCI", "HCI Design And Evaluation Methods", "Empirical Studies In HCI", "Interaction Paradigms", "Virtual Reality" ], "authors": [ { "affiliation": "Human-Computer Interaction University of Würzburg", "fullName": "Sebastian Oberdörfer", "givenName": "Sebastian", "surname": "Oberdörfer", "__typename": "ArticleAuthorType" }, { "affiliation": "Human-Computer Interaction University of Würzburg", "fullName": "Samantha Straka", "givenName": "Samantha", "surname": "Straka", "__typename": "ArticleAuthorType" }, { "affiliation": "Human-Computer Interaction University of Würzburg", "fullName": "Marc Erich Latoschik", "givenName": "Marc Erich", "surname": "Latoschik", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-03-01T00:00:00", "pubType": "proceedings", "pages": "440-441", "year": "2021", "issn": null, "isbn": "978-1-6654-4057-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "405700a438", "articleId": "1tnXuRl9EJi", "__typename": "AdjacentArticleType" }, "next": { "fno": "405700a442", "articleId": "1tnXmVHCLRe", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ngmast/2016/0949/0/07801465", "title": "Immersive Virtual Reality as a Supplement in the Rehabilitation Program of Post-Stroke Patients", "doi": null, "abstractUrl": "/proceedings-article/ngmast/2016/07801465/12OmNrMZpyR", "parentPublication": { "id": "proceedings/ngmast/2016/0949/0", "title": "2016 10th International Conference on Next-Generation Mobile Applications, Security and Technologies (NGMAST)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/case/2011/1732/0/06042427", "title": "Tele-operation of a mobile mining robot using a panoramic display: an exploration of operators sense of presence", "doi": null, "abstractUrl": "/proceedings-article/case/2011/06042427/12OmNroijfW", "parentPublication": { "id": "proceedings/case/2011/1732/0", "title": "2011 IEEE International Conference on Automation Science and Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dexa/2008/3299/0/3299a444", "title": "The Effectiveness of the Nature and the Type of Web Content", "doi": null, "abstractUrl": "/proceedings-article/dexa/2008/3299a444/12OmNxX3uAH", "parentPublication": { "id": "proceedings/dexa/2008/3299/0", "title": "2008 19th International Workshop on Database and Expert Systems Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2017/6327/0/6327a310", "title": "Improving the Immersion in Virtual Reality with Real-Time Avatar and Haptic Feedback in a Cricket Simulation", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2017/6327a310/12OmNyGbIfg", "parentPublication": { "id": "proceedings/ismar-adjunct/2017/6327/0", "title": "2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2017/6647/0/07892288", "title": "The impact of transitions on user experience in virtual reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892288/12OmNzUPptg", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446221", "title": "Simulator Sick but Still Immersed: A Comparison of Head-Object Collision Handling and Their Impact on Fun, Immersion, and Simulator Sickness", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446221/13bd1eSlytz", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446046", "title": "The Effect of Immersion on Emotional Responses to Film Viewing in a Virtual Environment", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446046/13bd1gCd7Th", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089482", "title": "Think Twice: The Influence of Immersion on Decision Making during Gambling in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089482/1jIxeHgBEkg", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2020/7675/0/767500a184", "title": "Effects of Augmented Content&#x2019;s Placement and Size on User&#x2019;s Search Experience in Extended Displays", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2020/767500a184/1pBMk7KgZHy", "parentPublication": { "id": "proceedings/ismar-adjunct/2020/7675/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2021/4057/0/405700a494", "title": "Indicators and Predictors of the Suspension of Disbelief: Children&#x2019;s Individual Presence Tendencies", "doi": null, "abstractUrl": "/proceedings-article/vrw/2021/405700a494/1tnWQ58g52g", "parentPublication": { "id": "proceedings/vrw/2021/4057/0", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyugyE7", "title": "Internet and Web Applications and Services, International Conference on", "acronym": "iciw", "groupId": "1001764", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNBPc8ub", "doi": "10.1109/ICIW.2008.21", "title": "Experiences with Virtual Environments in the EHEA", "normalizedTitle": "Experiences with Virtual Environments in the EHEA", "abstract": "Recent adaptations of educational curricula to the European Higher Education Area (EHEA) enabled many education professionals to adopt new teaching methodologies into their daily practice. Although the transmission of the main theoretic and practical concepts of the subject in question must remain the main focus, the improvement of students' general skills claims for special attention not only in the EHEA but also with a view to business requirements. Besides, the mobility of teachers and students among European universities has increased notably, which is a very positive experience for both communities. In this regard, new educational alternatives have to be found for this purpose, not only to provide the possibility of studying a subject abroad, but also to improve cross-curricular capabilities. It is also a reality to be taken into account that an increasing number of students are already working in the industry while following up their studies. E-learning environments are the perfect answer to these requirements, where a varied range of activities and methodologies can be used for allowing to follow the subject virtually as well as promote general cross-curricular skills such as addressing an audience, team work and preparing documentation. In this sense this paper describes the author's experiences with e-learning environments in the situations described above, which have been carried out with positive results: students have improved their general cross-curricular skills and were able to follow the subject from outside the university campus.", "abstracts": [ { "abstractType": "Regular", "content": "Recent adaptations of educational curricula to the European Higher Education Area (EHEA) enabled many education professionals to adopt new teaching methodologies into their daily practice. Although the transmission of the main theoretic and practical concepts of the subject in question must remain the main focus, the improvement of students' general skills claims for special attention not only in the EHEA but also with a view to business requirements. Besides, the mobility of teachers and students among European universities has increased notably, which is a very positive experience for both communities. In this regard, new educational alternatives have to be found for this purpose, not only to provide the possibility of studying a subject abroad, but also to improve cross-curricular capabilities. It is also a reality to be taken into account that an increasing number of students are already working in the industry while following up their studies. E-learning environments are the perfect answer to these requirements, where a varied range of activities and methodologies can be used for allowing to follow the subject virtually as well as promote general cross-curricular skills such as addressing an audience, team work and preparing documentation. In this sense this paper describes the author's experiences with e-learning environments in the situations described above, which have been carried out with positive results: students have improved their general cross-curricular skills and were able to follow the subject from outside the university campus.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Recent adaptations of educational curricula to the European Higher Education Area (EHEA) enabled many education professionals to adopt new teaching methodologies into their daily practice. Although the transmission of the main theoretic and practical concepts of the subject in question must remain the main focus, the improvement of students' general skills claims for special attention not only in the EHEA but also with a view to business requirements. Besides, the mobility of teachers and students among European universities has increased notably, which is a very positive experience for both communities. In this regard, new educational alternatives have to be found for this purpose, not only to provide the possibility of studying a subject abroad, but also to improve cross-curricular capabilities. It is also a reality to be taken into account that an increasing number of students are already working in the industry while following up their studies. E-learning environments are the perfect answer to these requirements, where a varied range of activities and methodologies can be used for allowing to follow the subject virtually as well as promote general cross-curricular skills such as addressing an audience, team work and preparing documentation. In this sense this paper describes the author's experiences with e-learning environments in the situations described above, which have been carried out with positive results: students have improved their general cross-curricular skills and were able to follow the subject from outside the university campus.", "fno": "3163a054", "keywords": [ "Virtual Environments", "EHEA", "Teaching Experience", "Cross Curricular Capacities", "Mobility Related Learning" ], "authors": [ { "affiliation": null, "fullName": "Guadalupe Ortiz", "givenName": "Guadalupe", "surname": "Ortiz", "__typename": "ArticleAuthorType" } ], "idPrefix": "iciw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-06-01T00:00:00", "pubType": "proceedings", "pages": "54-59", "year": "2008", "issn": null, "isbn": "978-0-7695-3163-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3163a049", "articleId": "12OmNqOwQH8", "__typename": "AdjacentArticleType" }, "next": { "fno": "04545608", "articleId": "141AnpJgoMC", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/fie/2014/3922/0/07044076", "title": "Co-curricular experiences link to nontechnical skill development for African-American engineers: Communication, teamwork, professionalism, lifelong learning, and reflective behavior skills", "doi": null, "abstractUrl": "/proceedings-article/fie/2014/07044076/12OmNAYGlpN", "parentPublication": { "id": "proceedings/fie/2014/3922/0", "title": "2014 IEEE Frontiers in Education Conference (FIE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/enc/2009/3882/0/3882a289", "title": "The Jigsaw Technique: Experiences Teaching Analysis Class Diagrams", "doi": null, "abstractUrl": "/proceedings-article/enc/2009/3882a289/12OmNBNM8Or", "parentPublication": { "id": "proceedings/enc/2009/3882/0", "title": "2009 Mexican International Conference on Computer Science", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fie/2011/468/0/06142752", "title": "Automatic assessment for the e-learning of the network services in the context of the EHEA", "doi": null, "abstractUrl": "/proceedings-article/fie/2011/06142752/12OmNqBbHXC", "parentPublication": { "id": "proceedings/fie/2011/468/0", "title": "2011 Frontiers in Education Conference (FIE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iciw/2009/3613/0/3613a539", "title": "Teaching Web Services with Virtual Environments: Improving Cross-Curricular Skills", "doi": null, "abstractUrl": "/proceedings-article/iciw/2009/3613a539/12OmNqJq4Eg", "parentPublication": { "id": "proceedings/iciw/2009/3613/0", "title": "Internet and Web Applications and Services, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fie/2010/6261/0/05673325", "title": "Developing civil engineering leaders at the United States Coast Guard Academy", "doi": null, "abstractUrl": "/proceedings-article/fie/2010/05673325/12OmNx4yvw7", "parentPublication": { "id": "proceedings/fie/2010/6261/0", "title": "2010 IEEE Frontiers in Education Conference (FIE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fie/2004/8552/0/01408627", "title": "Creating project-based learning experiences for University-K-12 partnerships", "doi": null, "abstractUrl": "/proceedings-article/fie/2004/01408627/12OmNyPQ4Ld", "parentPublication": { "id": "proceedings/fie/2004/8552/0", "title": "34th Annual Frontiers in Education, 2004. FIE 2004.", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fie/2014/3922/0/07044179", "title": "Improving communication skills: Students' viewpoint on a content & language integrated learning project", "doi": null, "abstractUrl": "/proceedings-article/fie/2014/07044179/12OmNyQ7FMs", "parentPublication": { "id": "proceedings/fie/2014/3922/0", "title": "2014 IEEE Frontiers in Education Conference (FIE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/1997/8076/0/80760242", "title": "From concept to image with computer art graphics", "doi": null, "abstractUrl": "/proceedings-article/iv/1997/80760242/12OmNyQpgVS", "parentPublication": { "id": "proceedings/iv/1997/8076/0", "title": "Proceedings. 1997 IEEE Conference on Information Visualization (Cat. No.97TB100165)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fie/2007/1083/0/04418179", "title": "Playing and learning? Educational gaming for engineering education", "doi": null, "abstractUrl": "/proceedings-article/fie/2007/04418179/12OmNzYNN0Q", "parentPublication": { "id": "proceedings/fie/2007/1083/0", "title": "2007 37th Annual Frontiers in Education Conference - Global Engineering: Knowledge Without Borders, Opportunities Without Passports", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cseet/2012/1592/0/06245008", "title": "Experiences with Integrating Simulation into a Software Engineering Curriculum", "doi": null, "abstractUrl": "/proceedings-article/cseet/2012/06245008/12OmNzaQody", "parentPublication": { "id": "proceedings/cseet/2012/1592/0", "title": "2012 IEEE 25th Conference on Software Engineering Education and Training (CSEE&T)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwnH4Qo", "title": "Proceedings of the 34th Annual Hawaii International Conference on System Sciences", "acronym": "hicss", "groupId": "1000730", "volume": "1", "displayVolume": "2", "year": "2001", "__typename": "ProceedingType" }, "article": { "id": "12OmNBlFQZa", "doi": "10.1109/HICSS.2001.926213", "title": "Sharing Viewpoints in Collaborative Virtual Environments", "normalizedTitle": "Sharing Viewpoints in Collaborative Virtual Environments", "abstract": "In this paper, we explore to what degree-shared viewpoints in three-dimensional collaborative virtual environments enable effective collaboration. The paper applies research on shared viewpoints and telepointers to 3D environments. A lightweight Java-based tool for creating collaborative virtual environments was developed and used in the study. The system is realized as an application framework that can be customized to develop new applications. We conducted a series of experiments to assess the effectiveness of shared viewpoints on two simple tasks. Control groups were provided with telepointers. Experimental groups were provided with telepointers and shared views. The results indicate that for participants with access to both tools, shared views are preferred over telepointers for tasks involving joint exploration of either the environment or some object of common interest.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, we explore to what degree-shared viewpoints in three-dimensional collaborative virtual environments enable effective collaboration. The paper applies research on shared viewpoints and telepointers to 3D environments. A lightweight Java-based tool for creating collaborative virtual environments was developed and used in the study. The system is realized as an application framework that can be customized to develop new applications. We conducted a series of experiments to assess the effectiveness of shared viewpoints on two simple tasks. Control groups were provided with telepointers. Experimental groups were provided with telepointers and shared views. The results indicate that for participants with access to both tools, shared views are preferred over telepointers for tasks involving joint exploration of either the environment or some object of common interest.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, we explore to what degree-shared viewpoints in three-dimensional collaborative virtual environments enable effective collaboration. The paper applies research on shared viewpoints and telepointers to 3D environments. A lightweight Java-based tool for creating collaborative virtual environments was developed and used in the study. The system is realized as an application framework that can be customized to develop new applications. We conducted a series of experiments to assess the effectiveness of shared viewpoints on two simple tasks. Control groups were provided with telepointers. Experimental groups were provided with telepointers and shared views. The results indicate that for participants with access to both tools, shared views are preferred over telepointers for tasks involving joint exploration of either the environment or some object of common interest.", "fno": "09811032", "keywords": [ "Collaborative Virtual Environments", "CSCW", "Groupware", "Viewpoint Sharing" ], "authors": [ { "affiliation": null, "fullName": "S. Valin", "givenName": "S.", "surname": "Valin", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "A. Francu", "givenName": "A.", "surname": "Francu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "H. Trefftz", "givenName": "H.", "surname": "Trefftz", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "I. Marsic", "givenName": "I.", "surname": "Marsic", "__typename": "ArticleAuthorType" } ], "idPrefix": "hicss", "isOpenAccess": true, "showRecommendedArticles": true, "showBuyMe": false, "hasPdf": true, "pubDate": "2001-01-01T00:00:00", "pubType": "proceedings", "pages": "1032", "year": "2001", "issn": "1530-1605", "isbn": "0-7695-0981-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09811031", "articleId": "12OmNyQYtu3", "__typename": "AdjacentArticleType" }, "next": { "fno": "09811033", "articleId": "12OmNyUWQVd", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/hicss/2001/0981/0/00926213", "title": "Sharing viewpoints in collaborative virtual environments", "doi": null, "abstractUrl": "/proceedings-article/hicss/2001/00926213/12OmNBf94YI", "parentPublication": { "id": "proceedings/hicss/2001/0981/2", "title": "Proceedings of the 34th Annual Hawaii International Conference on System Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csse/2008/3336/5/3336i125", "title": "A Virtual Group Workspace for Rationale Sharing in Distributed Collaborative Learning Activities", "doi": null, "abstractUrl": "/proceedings-article/csse/2008/3336i125/12OmNqIQSj9", "parentPublication": { "id": "proceedings/csse/2008/3336/5", "title": "2008 International Conference on Computer Science and Software Engineering (CSSE 2008)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1995/7084/0/70840148", "title": "Interacting in distributed collaborative virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vrais/1995/70840148/12OmNvAAto5", "parentPublication": { "id": "proceedings/vrais/1995/7084/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ozchi/1996/7525/0/75250054", "title": "Semantic Telepointers for Groupware", "doi": null, "abstractUrl": "/proceedings-article/ozchi/1996/75250054/12OmNvyjGgP", "parentPublication": { "id": "proceedings/ozchi/1996/7525/0", "title": "Proceedings Sixth Australian Conference on Computer-Human Interaction", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icgse-w/2011/4558/0/4558a009", "title": "Architectural Viewpoints for Global Software Development", "doi": null, "abstractUrl": "/proceedings-article/icgse-w/2011/4558a009/12OmNwDj12l", "parentPublication": { "id": "proceedings/icgse-w/2011/4558/0", "title": "Global Software Engineering Workshop, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sbsc/2012/4890/0/4890a035", "title": "Sanar: A Collaborative Environment to Support Knowledge Sharing with Medical Artifacts", "doi": null, "abstractUrl": "/proceedings-article/sbsc/2012/4890a035/12OmNwGIcBt", "parentPublication": { "id": "proceedings/sbsc/2012/4890/0", "title": "Sistemas Colaborativos, Simp&#x00F3;sio Brasilerio de", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vsmm/1997/8150/0/81500026", "title": "Collaborative Virtual Environments, Real-Time Video and Networking", "doi": null, "abstractUrl": "/proceedings-article/vsmm/1997/81500026/12OmNxFJXPy", "parentPublication": { "id": "proceedings/vsmm/1997/8150/0", "title": "Virtual Systems and MultiMedia, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmcs/1997/7819/0/78190466", "title": "Shared Media Space Coordination: Mixed Mode Synchrony in Collaborative Multimedia Environments", "doi": null, "abstractUrl": "/proceedings-article/icmcs/1997/78190466/12OmNxYbSZZ", "parentPublication": { "id": "proceedings/icmcs/1997/7819/0", "title": "Multimedia Computing and Systems, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2011/0039/0/05759478", "title": "Dual interactions between multi-display and smartphone for collaborative design and sharing", "doi": null, "abstractUrl": "/proceedings-article/vr/2011/05759478/12OmNzl3X0E", "parentPublication": { "id": "proceedings/vr/2011/0039/0", "title": "2011 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/1997/02/mcg1997020066", "title": "Visinet: Collaborative 3D Visualization and VR over ATM Networks", "doi": null, "abstractUrl": "/magazine/cg/1997/02/mcg1997020066/13rRUxlgxPb", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBqMDBD", "title": "Proceedings IEEE Virtual Reality 2002", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2002", "__typename": "ProceedingType" }, "article": { "id": "12OmNwHhoQ2", "doi": "10.1109/VR.2002.996536", "title": "Perceived Egocentric Distances in Real, Image-Based, and Traditional Virtual Environments", "normalizedTitle": "Perceived Egocentric Distances in Real, Image-Based, and Traditional Virtual Environments", "abstract": "In virtual environments, perceived egocentric distances are consistently underestimated when compared to the same distance judgments in the real world. The research presented in this paper explores two possible causes for the compressed distance perception in virtual environments: (1) real-time computer graphics rendering, and (2) immersive display technology. Our experiment compared absolute egocentric distance judgments in three complex, indoor environments: a real hallway with full-cue conditions; a virtual, stereoscopic, photographic panorama; and a virtual, stereoscopic computer model. Perceived egocentric distance was determined by a directed action task in which subjects walk blindfolded to a target. Our results show there is a significant difference in distance judgments between real and virtual environments. However, the differences between distance judgments in virtual photographic panorama environments and traditionally rendered virtual environments are small, suggesting that the display device is affecting distance judgments in virtual environments.", "abstracts": [ { "abstractType": "Regular", "content": "In virtual environments, perceived egocentric distances are consistently underestimated when compared to the same distance judgments in the real world. The research presented in this paper explores two possible causes for the compressed distance perception in virtual environments: (1) real-time computer graphics rendering, and (2) immersive display technology. Our experiment compared absolute egocentric distance judgments in three complex, indoor environments: a real hallway with full-cue conditions; a virtual, stereoscopic, photographic panorama; and a virtual, stereoscopic computer model. Perceived egocentric distance was determined by a directed action task in which subjects walk blindfolded to a target. Our results show there is a significant difference in distance judgments between real and virtual environments. However, the differences between distance judgments in virtual photographic panorama environments and traditionally rendered virtual environments are small, suggesting that the display device is affecting distance judgments in virtual environments.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In virtual environments, perceived egocentric distances are consistently underestimated when compared to the same distance judgments in the real world. The research presented in this paper explores two possible causes for the compressed distance perception in virtual environments: (1) real-time computer graphics rendering, and (2) immersive display technology. Our experiment compared absolute egocentric distance judgments in three complex, indoor environments: a real hallway with full-cue conditions; a virtual, stereoscopic, photographic panorama; and a virtual, stereoscopic computer model. Perceived egocentric distance was determined by a directed action task in which subjects walk blindfolded to a target. Our results show there is a significant difference in distance judgments between real and virtual environments. However, the differences between distance judgments in virtual photographic panorama environments and traditionally rendered virtual environments are small, suggesting that the display device is affecting distance judgments in virtual environments.", "fno": "14920275", "keywords": [ "Perception", "Virtual Environment", "Virtual Reality", "Immersive Environment", "Egocentric Distance" ], "authors": [ { "affiliation": "University of Utah", "fullName": "Peter Willemsen", "givenName": "Peter", "surname": "Willemsen", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Utah", "fullName": "Amy A. Gooch", "givenName": "Amy A.", "surname": "Gooch", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2002-03-01T00:00:00", "pubType": "proceedings", "pages": "275", "year": "2002", "issn": "1087-8270", "isbn": "0-7695-1492-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "14920273", "articleId": "12OmNrY3Luv", "__typename": "AdjacentArticleType" }, "next": { "fno": "14920277", "articleId": "12OmNx8wTr1", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2007/0905/0/04161000", "title": "Elucidating Factors that can Facilitate Veridical Spatial Perception in Immersive Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2007/04161000/12OmNAtaRZ9", "parentPublication": { "id": "proceedings/vr/2007/0905/0", "title": "2007 IEEE Virtual Reality Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2006/0224/0/02240003", "title": "Distance Perception in Immersive Virtual Environments, Revisited", "doi": null, "abstractUrl": "/proceedings-article/vr/2006/02240003/12OmNvm6VKz", "parentPublication": { "id": "proceedings/vr/2006/0224/0", "title": "IEEE Virtual Reality Conference (VR 2006)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444802", "title": "Avatar self-embodiment enhances distance perception accuracy in non-photorealistic immersive virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444802/12OmNwD1pSy", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2011/0039/0/05759475", "title": "Peripheral visual information and its effect on the perception of egocentric depth in virtual and augmented environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2011/05759475/12OmNySG3V8", "parentPublication": { "id": "proceedings/vr/2011/0039/0", "title": "2011 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444791", "title": "Influence of tactile feedback and presence on egocentric distance perception in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444791/12OmNyoAA64", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2000/6478/0/64780032", "title": "Visual Cues for Imminent Object Contact in Realistic Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2000/64780032/12OmNzIUfY6", "parentPublication": { "id": "proceedings/ieee-vis/2000/6478/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2011/0039/0/05759434", "title": "Self-motion illusions in immersive virtual reality environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2011/05759434/12OmNzZmZon", "parentPublication": { "id": "proceedings/vr/2011/0039/0", "title": "2011 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040589", "title": "Geometric Calibration of Head-Mounted Displays and its Effects on Distance Estimation", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040589/13rRUwbs2b1", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/03/v0429", "title": "Egocentric Depth Judgments in Optical, See-Through Augmented Reality", "doi": null, "abstractUrl": "/journal/tg/2007/03/v0429/13rRUxYrbM5", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2021/1838/0/255600a122", "title": "Augmented Reality for Maritime Navigation Assistance - Egocentric Depth Perception in Large Distance Outdoor Environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2021/255600a122/1tuB9Rs5D2M", "parentPublication": { "id": "proceedings/vr/2021/1838/0", "title": "2021 IEEE Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwdbV00", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNyKJinJ", "doi": "10.1109/CVPR.2012.6247820", "title": "Discovering important people and objects for egocentric video summarization", "normalizedTitle": "Discovering important people and objects for egocentric video summarization", "abstract": "We developed an approach to summarize egocentric video. We introduced novel egocentric features to train a regressor that predicts important regions. Using the discovered important regions, our approach produces significantly more informative summaries than traditional methods that often include irrelevant or redundant information.", "abstracts": [ { "abstractType": "Regular", "content": "We developed an approach to summarize egocentric video. We introduced novel egocentric features to train a regressor that predicts important regions. Using the discovered important regions, our approach produces significantly more informative summaries than traditional methods that often include irrelevant or redundant information.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We developed an approach to summarize egocentric video. We introduced novel egocentric features to train a regressor that predicts important regions. Using the discovered important regions, our approach produces significantly more informative summaries than traditional methods that often include irrelevant or redundant information.", "fno": "170P2A20", "keywords": [ "Video Signal Processing", "Feature Extraction", "Object Detection", "Important Region Discovery", "Egocentric Video Summarization", "Egocentric Features", "Informative Summaries", "Important Object Discovery", "Important People Discovery", "Cameras", "Visualization", "Predictive Models", "Image Color Analysis", "Training", "Context", "Humans" ], "authors": [ { "affiliation": "Univ. of Texas at Austin, Austin, TX, USA", "fullName": "J. Ghosh", "givenName": "J.", "surname": "Ghosh", "__typename": "ArticleAuthorType" }, { "affiliation": "Univ. of Texas at Austin, Austin, TX, USA", "fullName": "Yong Jae Lee", "givenName": null, "surname": "Yong Jae Lee", "__typename": "ArticleAuthorType" }, { "affiliation": "Univ. of Texas at Austin, Austin, TX, USA", "fullName": "K. Grauman", "givenName": "K.", "surname": "Grauman", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-06-01T00:00:00", "pubType": "proceedings", "pages": "1346-1353", "year": "2012", "issn": "1063-6919", "isbn": "978-1-4673-1226-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "169P2A19", "articleId": "12OmNB0X8xJ", "__typename": "AdjacentArticleType" }, "next": { "fno": "171P2A21", "articleId": "12OmNAle6m2", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2015/6964/0/07298836", "title": "Gaze-enabled egocentric video summarization via constrained submodular maximization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2015/07298836/12OmNAGw16V", "parentPublication": { "id": "proceedings/cvpr/2015/6964/0", "title": "2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2016/1552/0/07574681", "title": "Wearable social camera: Egocentric video summarization for social interaction", "doi": null, "abstractUrl": "/proceedings-article/icmew/2016/07574681/12OmNvAiSEF", "parentPublication": { "id": "proceedings/icmew/2016/1552/0", "title": "2016 IEEE International Conference on Multimedia & Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2015/9711/0/5720a443", "title": "Summarizing While Recording: Context-Based Highlight Detection for Egocentric Videos", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2015/5720a443/12OmNwpGgN2", "parentPublication": { "id": "proceedings/iccvw/2015/9711/0", "title": "2015 IEEE International Conference on Computer Vision Workshop (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391e525", "title": "Storyline Representation of Egocentric Videos with an Applications to Story-Based Search", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391e525/12OmNx8wTlB", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2011/4589/0/4589a319", "title": "Affective Video Summarization and Story Board Generation Using Pupillary Dilation and Eye Gaze", "doi": null, "abstractUrl": "/proceedings-article/ism/2011/4589a319/12OmNy4IF0d", "parentPublication": { "id": "proceedings/ism/2011/4589/0", "title": "2011 IEEE International Symposium on Multimedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2013/4989/0/4989c714", "title": "Story-Driven Summarization for Egocentric Video", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2013/4989c714/12OmNy6HQUi", "parentPublication": { "id": "proceedings/cvpr/2013/4989/0", "title": "2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2011/0394/0/05995444", "title": "Learning to recognize objects in egocentric activities", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2011/05995444/12OmNzcPAur", "parentPublication": { "id": "proceedings/cvpr/2011/0394/0", "title": "CVPR 2011", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2021/8808/0/09412512", "title": "Flow-guided Spatial Attention Tracking for Egocentric Activity Recognition", "doi": null, "abstractUrl": "/proceedings-article/icpr/2021/09412512/1tmhX9aKR7G", "parentPublication": { "id": "proceedings/icpr/2021/8808/0", "title": "2020 25th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2023/06/09562265", "title": "Generating Personalized Summaries of Day Long Egocentric Videos", "doi": null, "abstractUrl": "/journal/tp/2023/06/09562265/1xtOoRYTrG0", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyKJiak", "title": "2008 23rd IEEE/ACM International Conference on Automated Software Engineering", "acronym": "ase", "groupId": "1000064", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNyKa60O", "doi": "10.1109/ASE.2008.58", "title": "VCR: Virtual Capture and Replay for Performance Testing", "normalizedTitle": "VCR: Virtual Capture and Replay for Performance Testing", "abstract": "This paper proposes a novel approach to performance testing, called virtual capture and replay (VCR), that couples capture and replay techniques with the checkpointing capabilities provided by the latest virtualization technologies. VCR enables software performance testers to automatically take a snapshot of a running system when certain critical conditions are verified, and to later replay the scenario that led to those conditions. Several in-depth analyses can be separately carried out in the laboratory just by rewinding the captured scenario and replaying it using different probes and analysis tools.", "abstracts": [ { "abstractType": "Regular", "content": "This paper proposes a novel approach to performance testing, called virtual capture and replay (VCR), that couples capture and replay techniques with the checkpointing capabilities provided by the latest virtualization technologies. VCR enables software performance testers to automatically take a snapshot of a running system when certain critical conditions are verified, and to later replay the scenario that led to those conditions. Several in-depth analyses can be separately carried out in the laboratory just by rewinding the captured scenario and replaying it using different probes and analysis tools.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper proposes a novel approach to performance testing, called virtual capture and replay (VCR), that couples capture and replay techniques with the checkpointing capabilities provided by the latest virtualization technologies. VCR enables software performance testers to automatically take a snapshot of a running system when certain critical conditions are verified, and to later replay the scenario that led to those conditions. Several in-depth analyses can be separately carried out in the laboratory just by rewinding the captured scenario and replaying it using different probes and analysis tools.", "fno": "04639351", "keywords": [ "Virtualization Technologies", "VCR", "Virtual Capture", "Virtual Replay", "Software Performance Testing", "Checkpointing Capabilities" ], "authors": [ { "affiliation": "Consiglio Naz. delle Ric., Ist. di Scienza e Tecnol. dell'Inf., Pisa", "fullName": "A. Bertolino", "givenName": "A.", "surname": "Bertolino", "__typename": "ArticleAuthorType" }, { "affiliation": "Consiglio Naz. delle Ric., Ist. di Scienza e Tecnol. dell'Inf., Pisa", "fullName": "G. De Angelis", "givenName": "G.", "surname": "De Angelis", "__typename": "ArticleAuthorType" }, { "affiliation": "Consiglio Naz. delle Ric., Ist. di Scienza e Tecnol. dell'Inf., Pisa", "fullName": "A. Sabetta", "givenName": "A.", "surname": "Sabetta", "__typename": "ArticleAuthorType" } ], "idPrefix": "ase", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-09-01T00:00:00", "pubType": "proceedings", "pages": "399-402", "year": "2008", "issn": null, "isbn": "978-1-4244-2187-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "04639350", "articleId": "12OmNCwUmCL", "__typename": "AdjacentArticleType" }, "next": { "fno": "04639352", "articleId": "12OmNBOUxsc", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/hpcc-css-icess/2015/8937/0/07336319", "title": "A Non-intrusive, Platform-Independent Capture/Replay Test Automation System", "doi": null, "abstractUrl": "/proceedings-article/hpcc-css-icess/2015/07336319/12OmNARiM2r", "parentPublication": { "id": "proceedings/hpcc-css-icess/2015/8937/0", "title": "2015 IEEE 17th International Conference on High Performance Computing and Communications (HPCC), 2015 IEEE 7th International Symposium on Cyberspace Safety and Security (CSS) and 2015 IEEE 12th International Conf on Embedded Software and Systems (ICESS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icstw/2011/4345/0/4345a572", "title": "Combining Model-Based and Capture-Replay Testing Techniques of Graphical User Interfaces: An Industrial Approach", "doi": null, "abstractUrl": "/proceedings-article/icstw/2011/4345a572/12OmNBt3qkf", "parentPublication": { "id": "proceedings/icstw/2011/4345/0", "title": "Software Testing Verification and Validation Workshop, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/waina/2012/4652/0/4652a429", "title": "Deterministic Replay of Multithread Applications Using Virtual Machine", "doi": null, "abstractUrl": "/proceedings-article/waina/2012/4652a429/12OmNC8MsGN", "parentPublication": { "id": "proceedings/waina/2012/4652/0", "title": "2012 26th International Conference on Advanced Information Networking and Applications Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/camp/1995/7134/0/71340132", "title": "Design, implementation, and performance of a scalable multi-camera interactive video capture system", "doi": null, "abstractUrl": "/proceedings-article/camp/1995/71340132/12OmNwDj147", "parentPublication": { "id": "proceedings/camp/1995/7134/0", "title": "Computer Architectures for Machine Perception, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/compsac/2015/6564/3/6564c568", "title": "Replay Debugging of Real-Time Vxworks Applications", "doi": null, "abstractUrl": "/proceedings-article/compsac/2015/6564c568/12OmNwHhoUJ", "parentPublication": { "id": "proceedings/compsac/2015/6564/3", "title": "2015 IEEE 39th Annual Computer Software and Applications Conference (COMPSAC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fpl/2010/4179/0/4179a179", "title": "Secure Protocol Implementation for Remote Bitstream Update Preventing Replay Attacks on FPGA", "doi": null, "abstractUrl": "/proceedings-article/fpl/2010/4179a179/12OmNxd4tsU", "parentPublication": { "id": "proceedings/fpl/2010/4179/0", "title": "International Conference on Field Programmable Logic and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ds-rt/2008/3425/0/3425a252", "title": "A Tool for Replay and Analysis of Gaze-Enhanced Multiparty Sessions Captured in Immersive Collaborative Environments", "doi": null, "abstractUrl": "/proceedings-article/ds-rt/2008/3425a252/12OmNyYm2Bp", "parentPublication": { "id": "proceedings/ds-rt/2008/3425/0", "title": "Distributed Simulation and Real Time Applications, IEEE/ACM International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csmr/2012/4666/0/4666a553", "title": "Deterministic Replay of System's Execution with Multi-target QEMU Simulator for Dynamic Analysis and Reverse Debugging", "doi": null, "abstractUrl": "/proceedings-article/csmr/2012/4666a553/12OmNzYwbUt", "parentPublication": { "id": "proceedings/csmr/2012/4666/0", "title": "2012 16th European Conference on Software Maintenance and Reengineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/itng/2010/3984/0/3984b038", "title": "A Systematic Capture and Replay Strategy for Testing Complex GUI Based Java Applications", "doi": null, "abstractUrl": "/proceedings-article/itng/2010/3984b038/12OmNzdoN3q", "parentPublication": { "id": "proceedings/itng/2010/3984/0", "title": "Information Technology: New Generations, Third International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500a337", "title": "Knowledge Capture and Replay for Continual Learning", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500a337/1B13mlRu9gY", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzcxZq4", "title": "2009 International Conference on Signal Processing Systems (ICSPS)", "acronym": "icsps", "groupId": "1002804", "volume": "0", "displayVolume": "0", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNyXMQeH", "doi": "10.1109/ICSPS.2009.76", "title": "X3D-Based Virtual Reality Experiences in Water Museum Exhibitions", "normalizedTitle": "X3D-Based Virtual Reality Experiences in Water Museum Exhibitions", "abstract": "This paper discusses the work undertaken by the China Water Museum to develop X3D-based Water Museum Virtual Reality Experience System (WMVRES) to educate the public on water resources and the history of water. This system is to deal with difficulties in applying 3D information system in virtual reality experience, which is written in Extensible-3D(X3D) and Java. Administrators of this system manage information in 3D scene and the updated information can be saved in database automatically. Two key technologies are involved in this paper: Real-time rendering and Interaction. At the end of this paper, framework and major modules of the experience system are elaborated.", "abstracts": [ { "abstractType": "Regular", "content": "This paper discusses the work undertaken by the China Water Museum to develop X3D-based Water Museum Virtual Reality Experience System (WMVRES) to educate the public on water resources and the history of water. This system is to deal with difficulties in applying 3D information system in virtual reality experience, which is written in Extensible-3D(X3D) and Java. Administrators of this system manage information in 3D scene and the updated information can be saved in database automatically. Two key technologies are involved in this paper: Real-time rendering and Interaction. At the end of this paper, framework and major modules of the experience system are elaborated.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper discusses the work undertaken by the China Water Museum to develop X3D-based Water Museum Virtual Reality Experience System (WMVRES) to educate the public on water resources and the history of water. This system is to deal with difficulties in applying 3D information system in virtual reality experience, which is written in Extensible-3D(X3D) and Java. Administrators of this system manage information in 3D scene and the updated information can be saved in database automatically. Two key technologies are involved in this paper: Real-time rendering and Interaction. At the end of this paper, framework and major modules of the experience system are elaborated.", "fno": "3654a606", "keywords": [ "Virtual Reality Experience X 3 D Man Machine Interaction Graphics Rendering" ], "authors": [ { "affiliation": null, "fullName": "Chunlai Chai", "givenName": "Chunlai", "surname": "Chai", "__typename": "ArticleAuthorType" } ], "idPrefix": "icsps", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-05-01T00:00:00", "pubType": "proceedings", "pages": "606-609", "year": "2009", "issn": null, "isbn": "978-0-7695-3654-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3654a599", "articleId": "12OmNwl8GJC", "__typename": "AdjacentArticleType" }, "next": { "fno": "3654a610", "articleId": "12OmNxV4iyv", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iv-vis/2008/3271/0/3271a077", "title": "The Application of X3D to the Interactive Learning of Wave Optics", "doi": null, "abstractUrl": "/proceedings-article/iv-vis/2008/3271a077/12OmNAJ4pey", "parentPublication": { "id": "proceedings/iv-vis/2008/3271/0", "title": "Visualisation, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2009/3791/0/3791a036", "title": "Collaboration in 3D Shared Spaces Using X3D and VRML", "doi": null, "abstractUrl": "/proceedings-article/cw/2009/3791a036/12OmNvAAtmH", "parentPublication": { "id": "proceedings/cw/2009/3791/0", "title": "2009 International Conference on CyberWorlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dmdcm/2011/4413/0/4413a001", "title": "A Physics-Based Augmented Reality Jenga Stacking Game", "doi": null, "abstractUrl": "/proceedings-article/dmdcm/2011/4413a001/12OmNx5GTXh", "parentPublication": { "id": "proceedings/dmdcm/2011/4413/0", "title": "Digital Media and Digital Content Management, Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2016/1552/0/07574692", "title": "Creating new museum experiences for virtual reality", "doi": null, "abstractUrl": "/proceedings-article/icmew/2016/07574692/12OmNxVDuPe", "parentPublication": { "id": "proceedings/icmew/2016/1552/0", "title": "2016 IEEE International Conference on Multimedia & Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/jcdl/2004/832/0/01336227", "title": "Visiting virtual reality museum exhibits", "doi": null, "abstractUrl": "/proceedings-article/jcdl/2004/01336227/12OmNyTOssS", "parentPublication": { "id": "proceedings/jcdl/2004/832/0", "title": "Proceedings of the Fourth ACM/IEEE Joint Conference on Digital Libraries", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500c354", "title": "Auto-X3D: Ultra-Efficient Video Understanding via Finer-Grained Neural Architecture Search", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500c354/1B13o7f0NQk", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icipmc/2022/6872/0/687200a120", "title": "Application of Digital Audio-Visual Technology in Museum Exhibitions of Ancient Paintings&#x2014;Taking the Nanjing Museum as an Example", "doi": null, "abstractUrl": "/proceedings-article/icipmc/2022/687200a120/1GIura5rZBu", "parentPublication": { "id": "proceedings/icipmc/2022/6872/0", "title": "2022 International Conference on Image Processing and Media Computing (ICIPMC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aiars/2022/5457/0/545700a237", "title": "Realization of Digital Museum Design System Based on 3D Technology", "doi": null, "abstractUrl": "/proceedings-article/aiars/2022/545700a237/1J2XXMgp9qo", "parentPublication": { "id": "proceedings/aiars/2022/5457/0", "title": "2022 International Conference on Artificial Intelligence and Autonomous Robot Systems (AIARS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icalt/2019/3485/0/348500a357", "title": "Towards Enhancing User Experience through a Web-Based Augmented Reality Museum", "doi": null, "abstractUrl": "/proceedings-article/icalt/2019/348500a357/1cYi0fAR90A", "parentPublication": { "id": "proceedings/icalt/2019/3485/2161-377X", "title": "2019 IEEE 19th International Conference on Advanced Learning Technologies (ICALT)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2021/1298/0/129800a013", "title": "CubeMuseum: An Augmented Reality Prototype of Embodied Virtual Museum", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2021/129800a013/1yeQTz52lRS", "parentPublication": { "id": "proceedings/ismar-adjunct/2021/1298/0", "title": "2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzZmZqZ", "title": "2008 Canadian Conference on Computer and Robot Vision", "acronym": "crv", "groupId": "1001794", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNzRHONh", "doi": "10.1109/CRV.2008.39", "title": "Multiple View Integration and Display Using Virtual Mirrors", "normalizedTitle": "Multiple View Integration and Display Using Virtual Mirrors", "abstract": "This paper describes a technique, called V-Mirroring, for integrating videos taken from different cameras with different viewpoints of the same scene. The term V-Mirroring stems from the use of virtual mirrors in order to composite videos together. These mirrors are placed in the scene, near to the locations of the cameras. Thereafter, for any given camera, its corresponding video is overlaid with the remaining videos in the locations of the virtual mirrors closest to their respective cameras. Thus, the objects in the scene that are imaged by more than one camera, can then be viewed from multiple viewpoints in a single video. Previous approaches for compositing images or videos, such as panorama mosaicing, require that the input videos' image planes lie on the same, or approximately the same 2D plane, thereby losing the 3D feeling of the environment. In this work, videos can be taken from very different viewpoints and still be combined into a single video containing the differing videos.", "abstracts": [ { "abstractType": "Regular", "content": "This paper describes a technique, called V-Mirroring, for integrating videos taken from different cameras with different viewpoints of the same scene. The term V-Mirroring stems from the use of virtual mirrors in order to composite videos together. These mirrors are placed in the scene, near to the locations of the cameras. Thereafter, for any given camera, its corresponding video is overlaid with the remaining videos in the locations of the virtual mirrors closest to their respective cameras. Thus, the objects in the scene that are imaged by more than one camera, can then be viewed from multiple viewpoints in a single video. Previous approaches for compositing images or videos, such as panorama mosaicing, require that the input videos' image planes lie on the same, or approximately the same 2D plane, thereby losing the 3D feeling of the environment. In this work, videos can be taken from very different viewpoints and still be combined into a single video containing the differing videos.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper describes a technique, called V-Mirroring, for integrating videos taken from different cameras with different viewpoints of the same scene. The term V-Mirroring stems from the use of virtual mirrors in order to composite videos together. These mirrors are placed in the scene, near to the locations of the cameras. Thereafter, for any given camera, its corresponding video is overlaid with the remaining videos in the locations of the virtual mirrors closest to their respective cameras. Thus, the objects in the scene that are imaged by more than one camera, can then be viewed from multiple viewpoints in a single video. Previous approaches for compositing images or videos, such as panorama mosaicing, require that the input videos' image planes lie on the same, or approximately the same 2D plane, thereby losing the 3D feeling of the environment. In this work, videos can be taken from very different viewpoints and still be combined into a single video containing the differing videos.", "fno": "3153a286", "keywords": [ "View Integration", "Display", "Virtual Mirroring", "Video Compositing" ], "authors": [ { "affiliation": null, "fullName": "Carmen E. Au", "givenName": "Carmen E.", "surname": "Au", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "James J. Clark", "givenName": "James J.", "surname": "Clark", "__typename": "ArticleAuthorType" } ], "idPrefix": "crv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-05-01T00:00:00", "pubType": "proceedings", "pages": "286-293", "year": "2008", "issn": null, "isbn": "978-0-7695-3153-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3153a278", "articleId": "12OmNynJMEA", "__typename": "AdjacentArticleType" }, "next": { "fno": "3153a294", "articleId": "12OmNzxgHvc", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cgiv/2006/2606/0/26060220", "title": "Synthesizing Free-Viewpoint Images from Multiple View Videos in Soccer Stadium", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2006/26060220/12OmNB9bvjn", "parentPublication": { "id": "proceedings/cgiv/2006/2606/0", "title": "International Conference on Computer Graphics, Imaging and Visualisation (CGIV'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cso/2010/4030/2/4030b444", "title": "Virtual View Synthesis for Multi-view 3D Display", "doi": null, "abstractUrl": "/proceedings-article/cso/2010/4030b444/12OmNBkfRmx", "parentPublication": { "id": null, "title": null, "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2010/4109/0/4109a045", "title": "Multiple View Geometries for Mirrors and Cameras", "doi": null, "abstractUrl": "/proceedings-article/icpr/2010/4109a045/12OmNvA1hoj", "parentPublication": { "id": "proceedings/icpr/2010/4109/0", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/apcip/2009/3699/2/3699b031", "title": "3D Multi-view Autostereoscopic Display and Its Key Technologie", "doi": null, "abstractUrl": "/proceedings-article/apcip/2009/3699b031/12OmNvAS4pe", "parentPublication": { "id": "proceedings/apcip/2009/3699/1", "title": "Information Processing, Asia-Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dimpvt/2011/4369/0/4369a334", "title": "A Best-Next-View-Selection Algorithm for Multi-view Rendering", "doi": null, "abstractUrl": "/proceedings-article/3dimpvt/2011/4369a334/12OmNviHKiv", "parentPublication": { "id": "proceedings/3dimpvt/2011/4369/0", "title": "2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2005/9331/0/01521426", "title": "Free viewpoint video synthesis and presentation from multiple sporting videos", "doi": null, "abstractUrl": "/proceedings-article/icme/2005/01521426/12OmNwcCILL", "parentPublication": { "id": "proceedings/icme/2005/9331/0", "title": "2005 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icicta/2010/4077/1/4077a970", "title": "Color Correction for Multi-view Video Based on Color Variation Curve", "doi": null, "abstractUrl": "/proceedings-article/icicta/2010/4077a970/12OmNyRxFEF", "parentPublication": { "id": "proceedings/icicta/2010/4077/1", "title": "Intelligent Computation Technology and Automation, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2008/3358/0/3358a037", "title": "Synchronizing Video Cameras with Non-overlapping Fields of View", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2008/3358a037/12OmNz5apMq", "parentPublication": { "id": "proceedings/sibgrapi/2008/3358/0", "title": "2008 XXI Brazilian Symposium on Computer Graphics and Image Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/itng/2010/3984/0/3984b314", "title": "The Impact of View Spacing in Multi-view Video Compression Efficiency", "doi": null, "abstractUrl": "/proceedings-article/itng/2010/3984b314/12OmNzaQoc2", "parentPublication": { "id": "proceedings/itng/2010/3984/0", "title": "Information Technology: New Generations, Third International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2005/01/mcg2005010048", "title": "Lumisight Table: An Interactive View-Dependent Tabletop Display", "doi": null, "abstractUrl": "/magazine/cg/2005/01/mcg2005010048/13rRUxjQyxE", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqAU6sR", "title": "Image and Graphics, International Conference on", "acronym": "icig", "groupId": "1001790", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNzTH16l", "doi": "10.1109/ICIG.2011.43", "title": "Visual Saliency Based Aerial Video Summarization by Online Scene Classification", "normalizedTitle": "Visual Saliency Based Aerial Video Summarization by Online Scene Classification", "abstract": "Compared with traditional video summarization approaches, aerial video summarization is a new and challenging issue for its particular characteristics. Aerial video data is a massive data stream, without pre-edit structures such as sports or news video data, lack of camera motion such as zoom and pan. On account of these characteristics, we proposed a novel approach for summarization. First, we extract GIST features for each frame as the holistic scene representation. Then, we divide aerial video into temporal segments representing a visual scene using on-line clustering method by examine GIST features of each frame only once. Finally, we select several key frames from each scene for summarization according to visual saliency index (VSI) of each frame computed from their visual saliency map. In the paper, we proposed new criterion for estimation of temporal segmentation of streaming video. Experimental observations show the success of our approach on aerial video summarization.", "abstracts": [ { "abstractType": "Regular", "content": "Compared with traditional video summarization approaches, aerial video summarization is a new and challenging issue for its particular characteristics. Aerial video data is a massive data stream, without pre-edit structures such as sports or news video data, lack of camera motion such as zoom and pan. On account of these characteristics, we proposed a novel approach for summarization. First, we extract GIST features for each frame as the holistic scene representation. Then, we divide aerial video into temporal segments representing a visual scene using on-line clustering method by examine GIST features of each frame only once. Finally, we select several key frames from each scene for summarization according to visual saliency index (VSI) of each frame computed from their visual saliency map. In the paper, we proposed new criterion for estimation of temporal segmentation of streaming video. Experimental observations show the success of our approach on aerial video summarization.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Compared with traditional video summarization approaches, aerial video summarization is a new and challenging issue for its particular characteristics. Aerial video data is a massive data stream, without pre-edit structures such as sports or news video data, lack of camera motion such as zoom and pan. On account of these characteristics, we proposed a novel approach for summarization. First, we extract GIST features for each frame as the holistic scene representation. Then, we divide aerial video into temporal segments representing a visual scene using on-line clustering method by examine GIST features of each frame only once. Finally, we select several key frames from each scene for summarization according to visual saliency index (VSI) of each frame computed from their visual saliency map. In the paper, we proposed new criterion for estimation of temporal segmentation of streaming video. Experimental observations show the success of our approach on aerial video summarization.", "fno": "4541a777", "keywords": [ "Aerial Video Summarization", "Visual Attention", "Online Clustering", "Scene Classification", "Saliency" ], "authors": [ { "affiliation": null, "fullName": "Jiewei Wang", "givenName": "Jiewei", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yunhong Wang", "givenName": "Yunhong", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zhaoxiang Zhang", "givenName": "Zhaoxiang", "surname": "Zhang", "__typename": "ArticleAuthorType" } ], "idPrefix": "icig", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-08-01T00:00:00", "pubType": "proceedings", "pages": "777-782", "year": "2011", "issn": null, "isbn": "978-0-7695-4541-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4541a771", "articleId": "12OmNxWLTsQ", "__typename": "AdjacentArticleType" }, "next": { "fno": "4541a783", "articleId": "12OmNzlUKt8", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icig/2011/4541/0/4541a993", "title": "Aerial Video Images Registration Based on Optimal Derivative Filters with Scene-Adaptive Corners", "doi": null, "abstractUrl": "/proceedings-article/icig/2011/4541a993/12OmNApu5pP", "parentPublication": { "id": "proceedings/icig/2011/4541/0", "title": "Image and Graphics, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cmsp/2011/4356/1/4356a213", "title": "User-Specific Video Summarization", "doi": null, "abstractUrl": "/proceedings-article/cmsp/2011/4356a213/12OmNCcKQHO", "parentPublication": { "id": "proceedings/cmsp/2011/4356/1", "title": "Multimedia and Signal Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/das/2012/4661/0/4661a317", "title": "How Salient is Scene Text?", "doi": null, "abstractUrl": "/proceedings-article/das/2012/4661a317/12OmNqG0SQE", "parentPublication": { "id": "proceedings/das/2012/4661/0", "title": "Document Analysis Systems, IAPR International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cis/2011/4584/0/4584b188", "title": "Pedestrian Recognition Based on Saliency Detection and Kalman Filter Algorithm in Aerial Video", "doi": null, "abstractUrl": "/proceedings-article/cis/2011/4584b188/12OmNrIJqzj", "parentPublication": { "id": "proceedings/cis/2011/4584/0", "title": "2011 Seventh International Conference on Computational Intelligence and Security", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aipr/2012/4558/0/06528212", "title": "Spatial feature evaluation for aerial scene analysis", "doi": null, "abstractUrl": "/proceedings-article/aipr/2012/06528212/12OmNwfKjaI", "parentPublication": { "id": "proceedings/aipr/2012/4558/0", "title": "2012 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/isecs/2008/3258/0/3258a203", "title": "Automatic Video Summarization by Affinity Propagation Clustering and Semantic Content Mining", "doi": null, "abstractUrl": "/proceedings-article/isecs/2008/3258a203/12OmNxETadm", "parentPublication": { "id": "proceedings/isecs/2008/3258/0", "title": "Electronic Commerce and Security, International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2011/4589/0/4589a319", "title": "Affective Video Summarization and Story Board Generation Using Pupillary Dilation and Eye Gaze", "doi": null, "abstractUrl": "/proceedings-article/ism/2011/4589a319/12OmNy4IF0d", "parentPublication": { "id": "proceedings/ism/2011/4589/0", "title": "2011 IEEE International Symposium on Multimedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2007/1630/0/04408863", "title": "Scene Summarization for Online Image Collections", "doi": null, "abstractUrl": "/proceedings-article/iccv/2007/04408863/12OmNywfKEO", "parentPublication": { "id": "proceedings/iccv/2007/1630/0", "title": "2007 11th IEEE International Conference on Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/1998/11/i1254", "title": "A Model of Saliency-Based Visual Attention for Rapid Scene Analysis", "doi": null, "abstractUrl": "/journal/tp/1998/11/i1254/13rRUxcbnI9", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2019/5023/0/502300b545", "title": "Video Summarization by Learning Relationships between Action and Scene", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2019/502300b545/1i5mpUpYcso", "parentPublication": { "id": "proceedings/iccvw/2019/5023/0", "title": "2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwDAC3U", "title": "International Conference on Advances in Computer-Human Interaction", "acronym": "achi", "groupId": "1001548", "volume": "0", "displayVolume": "0", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNAZOJU1", "doi": "10.1109/ACHI.2009.61", "title": "Haptic Palpation for the Femoral Pulse in Virtual Interventional Radiology", "normalizedTitle": "Haptic Palpation for the Femoral Pulse in Virtual Interventional Radiology", "abstract": "Interventional Radiology is a rapidly expanding speciality using minimally invasive techniques to treat a multitude of clinical problems. Current work in progress aims to create an affordable virtual training tool to reduce training times and patient risk during a trainee practitioners learning cycle. The procedure of arterial catheterisation has been broken down into a number of subtasks, one of which requires an operator to locate the femoral artery pulse by palpation. This is performed in preparation for a needle insertion to allow the entry of a guide wire and catheter into the patient. This paper presents the current state of research into a unique solution for affordable haptic simulation of pulse palpation in a virtual environment.", "abstracts": [ { "abstractType": "Regular", "content": "Interventional Radiology is a rapidly expanding speciality using minimally invasive techniques to treat a multitude of clinical problems. Current work in progress aims to create an affordable virtual training tool to reduce training times and patient risk during a trainee practitioners learning cycle. The procedure of arterial catheterisation has been broken down into a number of subtasks, one of which requires an operator to locate the femoral artery pulse by palpation. This is performed in preparation for a needle insertion to allow the entry of a guide wire and catheter into the patient. This paper presents the current state of research into a unique solution for affordable haptic simulation of pulse palpation in a virtual environment.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Interventional Radiology is a rapidly expanding speciality using minimally invasive techniques to treat a multitude of clinical problems. Current work in progress aims to create an affordable virtual training tool to reduce training times and patient risk during a trainee practitioners learning cycle. The procedure of arterial catheterisation has been broken down into a number of subtasks, one of which requires an operator to locate the femoral artery pulse by palpation. This is performed in preparation for a needle insertion to allow the entry of a guide wire and catheter into the patient. This paper presents the current state of research into a unique solution for affordable haptic simulation of pulse palpation in a virtual environment.", "fno": "3529a193", "keywords": [ "Interventional Radiology", "Training", "Palpation", "Femoral Artery", "Simulation" ], "authors": [ { "affiliation": null, "fullName": "T. Coles", "givenName": "T.", "surname": "Coles", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "N. W. John", "givenName": "N. W.", "surname": "John", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "D. A. Gould", "givenName": "D. A.", "surname": "Gould", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "D. G. Caldwell", "givenName": "D. G.", "surname": "Caldwell", "__typename": "ArticleAuthorType" } ], "idPrefix": "achi", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-02-01T00:00:00", "pubType": "proceedings", "pages": "193-198", "year": "2009", "issn": null, "isbn": "978-0-7695-3529-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3529a181", "articleId": "12OmNrJAegF", "__typename": "AdjacentArticleType" }, "next": { "fno": "3529a199", "articleId": "12OmNxveNRn", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/svr/2014/4261/0/4261a268", "title": "Breast Palpation Simulation with Haptic Feedback: Prototype and Initial Results", "doi": null, "abstractUrl": "/proceedings-article/svr/2014/4261a268/12OmNBhHt7R", "parentPublication": { "id": "proceedings/svr/2014/4261/0", "title": "2014 XVI Symposium on Virtual and Augmented Reality (SVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cbms/1992/2742/0/00245023", "title": "Adaptive pulse rate scheduling for reduced dose X-ray cardiac interventional fluoroscopic procedures", "doi": null, "abstractUrl": "/proceedings-article/cbms/1992/00245023/12OmNrGKeuy", "parentPublication": { "id": "proceedings/cbms/1992/2742/0", "title": "Proceedings Fifth Annual IEEE Symposium on Computer-Based Medical Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/haptics/2008/2005/0/04479994", "title": "Force-Feedback Surgical Teleoperator: Controller Design and Palpation Experiments", "doi": null, "abstractUrl": "/proceedings-article/haptics/2008/04479994/12OmNvqW6Vk", "parentPublication": { "id": "proceedings/haptics/2008/2005/0", "title": "IEEE Haptics Symposium 2008", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/scamc/1983/0503/0/00764616", "title": "Managing a radiology department: a three dimensional problem", "doi": null, "abstractUrl": "/proceedings-article/scamc/1983/00764616/12OmNyRg4qC", "parentPublication": { "id": "proceedings/scamc/1983/0503/0", "title": "1983 The Seventh Annual Symposium on Computer Applications in Medical Care", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2011/03/tth2011030199", "title": "Integrating Haptics with Augmented Reality in a Femoral Palpation and Needle Insertion Training Simulation", "doi": null, "abstractUrl": "/journal/th/2011/03/tth2011030199/13rRUxd2aZb", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040617", "title": "Haptic Palpation for Medical Simulation in Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040617/13rRUyfKIHI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ichi/2022/6845/0/684500a090", "title": "Aspect-based Sentiment Analysis of Radiology Patient Experience Surveys: A Cohort Study", "doi": null, "abstractUrl": "/proceedings-article/ichi/2022/684500a090/1GvdtJrngf6", "parentPublication": { "id": "proceedings/ichi/2022/6845/0", "title": "2022 IEEE 10th International Conference on Healthcare Informatics (ICHI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNynsbxj", "title": "2014 XVI Symposium on Virtual and Augmented Reality (SVR)", "acronym": "svr", "groupId": "1800426", "volume": "0", "displayVolume": "0", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNBhHt7R", "doi": "10.1109/SVR.2014.57", "title": "Breast Palpation Simulation with Haptic Feedback: Prototype and Initial Results", "normalizedTitle": "Breast Palpation Simulation with Haptic Feedback: Prototype and Initial Results", "abstract": "Breast palpation, when a health professional touches the breast of a patient searching for nodules is an important screening test for the breast cancer detection. This exam consists of a main means of the early detection, however, many medical professionals do not have the suitable training to do so. This stems from the fact that exam training depends on volunteers who do not represent the real situation of the disease, or models restricted to a few cases, which do not reflect the range of characteristics of the cancer training. Because of this, this procedure has a high demand for simulators that are able to provide greater realism with low cost and flexibility. This article presents the definition and implementation of a breast palpation simulator with haptic feedback. The main contribution of this paper is to evaluate the haptic feedback provided by the simulator. Evaluation results of the tool from an experiment conducted by palpation with lay users are presented. From the results obtained, it was concluded that the simulator can generate significant learning for users with no experience in the area, with satisfactory force feedback.", "abstracts": [ { "abstractType": "Regular", "content": "Breast palpation, when a health professional touches the breast of a patient searching for nodules is an important screening test for the breast cancer detection. This exam consists of a main means of the early detection, however, many medical professionals do not have the suitable training to do so. This stems from the fact that exam training depends on volunteers who do not represent the real situation of the disease, or models restricted to a few cases, which do not reflect the range of characteristics of the cancer training. Because of this, this procedure has a high demand for simulators that are able to provide greater realism with low cost and flexibility. This article presents the definition and implementation of a breast palpation simulator with haptic feedback. The main contribution of this paper is to evaluate the haptic feedback provided by the simulator. Evaluation results of the tool from an experiment conducted by palpation with lay users are presented. From the results obtained, it was concluded that the simulator can generate significant learning for users with no experience in the area, with satisfactory force feedback.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Breast palpation, when a health professional touches the breast of a patient searching for nodules is an important screening test for the breast cancer detection. This exam consists of a main means of the early detection, however, many medical professionals do not have the suitable training to do so. This stems from the fact that exam training depends on volunteers who do not represent the real situation of the disease, or models restricted to a few cases, which do not reflect the range of characteristics of the cancer training. Because of this, this procedure has a high demand for simulators that are able to provide greater realism with low cost and flexibility. This article presents the definition and implementation of a breast palpation simulator with haptic feedback. The main contribution of this paper is to evaluate the haptic feedback provided by the simulator. Evaluation results of the tool from an experiment conducted by palpation with lay users are presented. From the results obtained, it was concluded that the simulator can generate significant learning for users with no experience in the area, with satisfactory force feedback.", "fno": "4261a268", "keywords": [ "Biomedical Education", "Cancer", "Computer Aided Instruction", "Digital Simulation", "Force Feedback", "Haptic Interfaces", "Medical Computing", "Breast Palpation Simulation", "Haptic Feedback", "Breast Nodules", "Breast Screening Test", "Breast Cancer Detection", "Medical Professional Training", "Cancer Training", "Force Feedback", "Haptic Interfaces", "Phantoms", "Java", "Breast", "Training", "Visualization", "Art", "Breast Palpation", "Simulation", "Haptics", "Force Feedback" ], "authors": [ { "affiliation": "Lab. de Aplic. de Inf. em Saude (LApIS), Univ. de Sao Paulo, Sao Paulo, Brazil", "fullName": "Mateus de Lara Ribeiro", "givenName": "Mateus", "surname": "de Lara Ribeiro", "__typename": "ArticleAuthorType" }, { "affiliation": "Lab. de Aplic. de Inf. em Saude (LApIS), Univ. de Sao Paulo, Sao Paulo, Brazil", "fullName": "Fátima L.S. Nunes", "givenName": "Fátima L.S.", "surname": "Nunes", "__typename": "ArticleAuthorType" } ], "idPrefix": "svr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-05-01T00:00:00", "pubType": "proceedings", "pages": "268-276", "year": "2014", "issn": null, "isbn": "978-1-4799-4261-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4261a258", "articleId": "12OmNzXnNpJ", "__typename": "AdjacentArticleType" }, "next": { "fno": "4261a277", "articleId": "12OmNrkT7qY", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2008/1971/0/04480757", "title": "Virtual Human + Tangible Interface = Mixed Reality Human An Initial Exploration with a Virtual Breast Exam Patient", "doi": null, "abstractUrl": "/proceedings-article/vr/2008/04480757/12OmNrJAdRV", "parentPublication": { "id": "proceedings/vr/2008/1971/0", "title": "IEEE Virtual Reality 2008", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2010/9343/0/05643617", "title": "Breast cancer palpation system using haptic augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2010/05643617/12OmNwF0C53", "parentPublication": { "id": "proceedings/ismar/2010/9343/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/waina/2015/1775/0/1775a044", "title": "Development of Virtual Palpation System for Dental Education", "doi": null, "abstractUrl": "/proceedings-article/waina/2015/1775a044/12OmNwMFMfZ", "parentPublication": { "id": "proceedings/waina/2015/1775/0", "title": "2015 IEEE 29th International Conference on Advanced Information Networking and Applications Workshops (WAINA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2010/9343/0/05643585", "title": "Haptic simulation of breast cancer palpation: A case study of haptic augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2010/05643585/12OmNwtn3ui", "parentPublication": { "id": "proceedings/ismar/2010/9343/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cbms/2016/9036/0/9036a279", "title": "Towards Determining Force Feedback Parameters for Realistic Representation of Nodules in a Breast Palpation Simulator", "doi": null, "abstractUrl": "/proceedings-article/cbms/2016/9036a279/12OmNx7G64H", "parentPublication": { "id": "proceedings/cbms/2016/9036/0", "title": "2016 IEEE 29th International Symposium on Computer-Based Medical Systems (CBMS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cbms/2011/1189/0/05999154", "title": "Ultrasound palpation by haptic elastography", "doi": null, "abstractUrl": "/proceedings-article/cbms/2011/05999154/12OmNy3Agyv", "parentPublication": { "id": "proceedings/cbms/2011/1189/0", "title": "2011 24th International Symposium on Computer-Based Medical Systems (CBMS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2014/4261/0/4261a258", "title": "Systematically Reviewing Techniques and Devices Used in Palpation Simulation with Haptic Feedback", "doi": null, "abstractUrl": "/proceedings-article/svr/2014/4261a258/12OmNzXnNpJ", "parentPublication": { "id": "proceedings/svr/2014/4261/0", "title": "2014 XVI Symposium on Virtual and Augmented Reality (SVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2012/02/tth2012020097", "title": "Abdominal Palpation Haptic Device for Colonoscopy Simulation Using Pneumatic Control", "doi": null, "abstractUrl": "/journal/th/2012/02/tth2012020097/13rRUwInvyJ", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040617", "title": "Haptic Palpation for Medical Simulation in Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040617/13rRUyfKIHI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aivr/2021/3225/0/322500a224", "title": "An Augmented Reality Application for Clinical Breast Examination Training", "doi": null, "abstractUrl": "/proceedings-article/aivr/2021/322500a224/1zxLvulckdq", "parentPublication": { "id": "proceedings/aivr/2021/3225/0", "title": "2021 IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwLOYSu", "title": "2017 International Conference on Cyberworlds (CW)", "acronym": "cw", "groupId": "1000175", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNCbU2U1", "doi": "10.1109/CW.2017.43", "title": "Modeling Deformable Objects for Medical Training with Haptic Devices", "normalizedTitle": "Modeling Deformable Objects for Medical Training with Haptic Devices", "abstract": "Medical training systems often require the use of haptic interaction with deformable models and high visual fidelity. It is necessary to get a compromise between the geometric model and the deformation model in order to achieve the required frame rate for haptic interaction of 1000Hz. Frequently low level geometric models are used to describe the geometry of the object. This produces large data structures for which it is difficult to reach real-time force feedback.This paper proposes to use a novel model based on Bezier hyperpatches to represent the object, computing the deformation using a mass-spring system. This model can represent free form objects in a concise way. A small test application to simulate palpation with a haptic probe is presented.", "abstracts": [ { "abstractType": "Regular", "content": "Medical training systems often require the use of haptic interaction with deformable models and high visual fidelity. It is necessary to get a compromise between the geometric model and the deformation model in order to achieve the required frame rate for haptic interaction of 1000Hz. Frequently low level geometric models are used to describe the geometry of the object. This produces large data structures for which it is difficult to reach real-time force feedback.This paper proposes to use a novel model based on Bezier hyperpatches to represent the object, computing the deformation using a mass-spring system. This model can represent free form objects in a concise way. A small test application to simulate palpation with a haptic probe is presented.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Medical training systems often require the use of haptic interaction with deformable models and high visual fidelity. It is necessary to get a compromise between the geometric model and the deformation model in order to achieve the required frame rate for haptic interaction of 1000Hz. Frequently low level geometric models are used to describe the geometry of the object. This produces large data structures for which it is difficult to reach real-time force feedback.This paper proposes to use a novel model based on Bezier hyperpatches to represent the object, computing the deformation using a mass-spring system. This model can represent free form objects in a concise way. A small test application to simulate palpation with a haptic probe is presented.", "fno": "2089a202", "keywords": [ "Computational Geometry", "Computer Based Training", "Force Feedback", "Haptic Interfaces", "Medical Computing", "Object Recognition", "Haptic Devices", "Medical Training Systems", "Haptic Interaction", "Geometric Model", "Deformation Model", "Mass Spring System", "Free Form Objects", "Haptic Probe", "Visual Fidelity", "Deformable Object Modeling", "Haptic Interfaces", "Computational Modeling", "Deformable Models", "Springs", "Strain", "Force", "Solid Modeling", "Deformable", "Medical", "Haptic", "Hyperpatch" ], "authors": [ { "affiliation": null, "fullName": "Celia Romo", "givenName": "Celia", "surname": "Romo", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Francisco A. Conde", "givenName": "Francisco A.", "surname": "Conde", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Nigel W. John", "givenName": "Nigel W.", "surname": "John", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Juan Carlos Torres", "givenName": "Juan Carlos", "surname": "Torres", "__typename": "ArticleAuthorType" } ], "idPrefix": "cw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-09-01T00:00:00", "pubType": "proceedings", "pages": "202-205", "year": "2017", "issn": null, "isbn": "978-1-5386-2089-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "2089a198", "articleId": "12OmNrYlmCA", "__typename": "AdjacentArticleType" }, "next": { "fno": "2089a206", "articleId": "12OmNwDACjB", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/haptic/2006/0226/0/01627099", "title": "A Displacement Driven Real-Time Deformable Model For Haptic Surgery Simulation", "doi": null, "abstractUrl": "/proceedings-article/haptic/2006/01627099/12OmNAXxWYS", "parentPublication": { "id": "proceedings/haptic/2006/0226/0", "title": "Haptic Interfaces for Virtual Environment and Teleoperator Systems, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/haptics/2010/6821/0/05444618", "title": "Physically-based analytical modelling of deformable haptic environments", "doi": null, "abstractUrl": "/proceedings-article/haptics/2010/05444618/12OmNBpVQ8r", "parentPublication": { "id": "proceedings/haptics/2010/6821/0", "title": "2010 IEEE Haptics Symposium (Formerly known as Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/haptic/2006/0226/0/01627118", "title": "Geometric Properties of Contacts Involving a Deformable Object", "doi": null, "abstractUrl": "/proceedings-article/haptic/2006/01627118/12OmNvTTc8v", "parentPublication": { "id": "proceedings/haptic/2006/0226/0", "title": "Haptic Interfaces for Virtual Environment and Teleoperator Systems, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/haptics/2006/0226/0/02260072", "title": "Physics-based s-Adaptive Haptic Simulation for Deformable Object", "doi": null, "abstractUrl": "/proceedings-article/haptics/2006/02260072/12OmNwE9OPe", "parentPublication": { "id": "proceedings/haptics/2006/0226/0", "title": "2006 14th Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2014/4717/0/06890578", "title": "Towards real-time modeling and haptic rendering of deformable objects for point cloud-based Model-Mediated Teleoperation", "doi": null, "abstractUrl": "/proceedings-article/icmew/2014/06890578/12OmNwJgALS", "parentPublication": { "id": "proceedings/icmew/2014/4717/0", "title": "2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2010/9343/0/05643585", "title": "Haptic simulation of breast cancer palpation: A case study of haptic augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2010/05643585/12OmNwtn3ui", "parentPublication": { "id": "proceedings/ismar/2010/9343/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/haptics/2006/0226/0/01627099", "title": "A Displacement Driven Real-Time Deformable Model For Haptic Surgery Simulation", "doi": null, "abstractUrl": "/proceedings-article/haptics/2006/01627099/12OmNx1IwaC", "parentPublication": { "id": "proceedings/haptics/2006/0226/0", "title": "2006 14th Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/haptic/2006/0226/0/01627112", "title": "Physics-based s-Adaptive Haptic Simulation for Deformable Object", "doi": null, "abstractUrl": "/proceedings-article/haptic/2006/01627112/12OmNzG4gug", "parentPublication": { "id": "proceedings/haptic/2006/0226/0", "title": "Haptic Interfaces for Virtual Environment and Teleoperator Systems, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2001/0948/0/09480264", "title": "Realistic Haptic Rendering for Highly Deformable Virtual Objects", "doi": null, "abstractUrl": "/proceedings-article/vr/2001/09480264/12OmNzJbQXR", "parentPublication": { "id": "proceedings/vr/2001/0948/0", "title": "Virtual Reality Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2016/04/07476889", "title": "Data-Driven Haptic Modeling and Rendering of Viscoelastic and Frictional Responses of Deformable Objects", "doi": null, "abstractUrl": "/journal/th/2016/04/07476889/13rRUwfZBVt", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyKa6fd", "title": "2016 10th International Conference on Complex, Intelligent, and Software Intensive Systems (CISIS)", "acronym": "cisis", "groupId": "1002005", "volume": "0", "displayVolume": "0", "year": "2016", "__typename": "ProceedingType" }, "article": { "id": "12OmNxVDuQI", "doi": "10.1109/CISIS.2016.142", "title": "Improvement of Mesh Free Deforming Analysis for Maxillofacial Palpation on a Virtual Training System", "normalizedTitle": "Improvement of Mesh Free Deforming Analysis for Maxillofacial Palpation on a Virtual Training System", "abstract": "Palpation is important medical diagnostic technic to make correct diagnosis of symptomatic state. It requires a lot of palpation experience under leadership of an experienced medical doctor. Hence, a palpation training system with virtual model of an affected part is useful to get better at the skill. However, there is no practical training for maxillofacial palpation in the dental education curriculum in Japan. So, it is necessary for dental students that the virtual training system is developed, which can to practice maxillofacial palpation. Then, a trial palpation training system for maxillofacial disorder is developed by using virtual reality model and haptic interface. The system is consisted of a PC and the haptic device - PHANToM Omni. A virtual affected part model is constructed from multi-CT images through commercial software. In particular, the dynamic responding virtual model and real-time analysis method through simplified mesh free method are improved. Consequently, we succeed in operation at the real-time by the palpation training system.", "abstracts": [ { "abstractType": "Regular", "content": "Palpation is important medical diagnostic technic to make correct diagnosis of symptomatic state. It requires a lot of palpation experience under leadership of an experienced medical doctor. Hence, a palpation training system with virtual model of an affected part is useful to get better at the skill. However, there is no practical training for maxillofacial palpation in the dental education curriculum in Japan. So, it is necessary for dental students that the virtual training system is developed, which can to practice maxillofacial palpation. Then, a trial palpation training system for maxillofacial disorder is developed by using virtual reality model and haptic interface. The system is consisted of a PC and the haptic device - PHANToM Omni. A virtual affected part model is constructed from multi-CT images through commercial software. In particular, the dynamic responding virtual model and real-time analysis method through simplified mesh free method are improved. Consequently, we succeed in operation at the real-time by the palpation training system.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Palpation is important medical diagnostic technic to make correct diagnosis of symptomatic state. It requires a lot of palpation experience under leadership of an experienced medical doctor. Hence, a palpation training system with virtual model of an affected part is useful to get better at the skill. However, there is no practical training for maxillofacial palpation in the dental education curriculum in Japan. So, it is necessary for dental students that the virtual training system is developed, which can to practice maxillofacial palpation. Then, a trial palpation training system for maxillofacial disorder is developed by using virtual reality model and haptic interface. The system is consisted of a PC and the haptic device - PHANToM Omni. A virtual affected part model is constructed from multi-CT images through commercial software. In particular, the dynamic responding virtual model and real-time analysis method through simplified mesh free method are improved. Consequently, we succeed in operation at the real-time by the palpation training system.", "fno": "0987a382", "keywords": [ "Training", "Computational Modeling", "Finite Element Analysis", "Dentistry", "Haptic Interfaces", "Biomedical Imaging", "Analytical Models", "Medical Simulator", "Virtual Reality", "Palpation Training System", "Force Sensation", "Linear Mesh Free Method" ], "authors": [ { "affiliation": null, "fullName": "Kenji Yoshitomi", "givenName": "Kenji", "surname": "Yoshitomi", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Tatsushi Tokuyasu", "givenName": "Tatsushi", "surname": "Tokuyasu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Kazuhiko Toshimitsu", "givenName": "Kazuhiko", "surname": "Toshimitsu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Takaaki Nakayama", "givenName": "Takaaki", "surname": "Nakayama", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Kazutoshi Okamura", "givenName": "Kazutoshi", "surname": "Okamura", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Kazunori Yoshiura", "givenName": "Kazunori", "surname": "Yoshiura", "__typename": "ArticleAuthorType" } ], "idPrefix": "cisis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2016-07-01T00:00:00", "pubType": "proceedings", "pages": "382-384", "year": "2016", "issn": null, "isbn": "978-1-5090-0987-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0987a377", "articleId": "12OmNBJw9Pt", "__typename": "AdjacentArticleType" }, "next": { "fno": "0987a385", "articleId": "12OmNrHjqKO", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/itme/2016/3906/0/3906a424", "title": "Application of a 3D Haptic Virtual Reality Simulation System for Dental Crown Preparation Training", "doi": null, "abstractUrl": "/proceedings-article/itme/2016/3906a424/12OmNrAMEPv", "parentPublication": { "id": "proceedings/itme/2016/3906/0", "title": "2016 8th International Conference on Information Technology in Medicine and Education (ITME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/waina/2015/1775/0/1775a044", "title": "Development of Virtual Palpation System for Dental Education", "doi": null, "abstractUrl": "/proceedings-article/waina/2015/1775a044/12OmNwMFMfZ", "parentPublication": { "id": "proceedings/waina/2015/1775/0", "title": "2015 IEEE 29th International Conference on Advanced Information Networking and Applications Workshops (WAINA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2010/9343/0/05643585", "title": "Haptic simulation of breast cancer palpation: A case study of haptic augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2010/05643585/12OmNwtn3ui", "parentPublication": { "id": "proceedings/ismar/2010/9343/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2013/5001/0/06655760", "title": "Haptic Rendering Techniques Supporting Material Removal", "doi": null, "abstractUrl": "/proceedings-article/svr/2013/06655760/12OmNz6iOwZ", "parentPublication": { "id": "proceedings/svr/2013/5001/0", "title": "2013 XV Symposium on Virtual and Augmented Reality (SVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2012/4814/0/4814a207", "title": "Virtual Palpation for Medical Training in Cyberworlds", "doi": null, "abstractUrl": "/proceedings-article/cw/2012/4814a207/12OmNzFv4gM", "parentPublication": { "id": "proceedings/cw/2012/4814/0", "title": "2012 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2014/4261/0/4261a258", "title": "Systematically Reviewing Techniques and Devices Used in Palpation Simulation with Haptic Feedback", "doi": null, "abstractUrl": "/proceedings-article/svr/2014/4261a258/12OmNzXnNpJ", "parentPublication": { "id": "proceedings/svr/2014/4261/0", "title": "2014 XVI Symposium on Virtual and Augmented Reality (SVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2011/01/tth2011010039", "title": "Physics-Based Haptic Simulation of Bone Machining", "doi": null, "abstractUrl": "/journal/th/2011/01/tth2011010039/13rRUwIF6le", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040617", "title": "Haptic Palpation for Medical Simulation in Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040617/13rRUyfKIHI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNB8Cj8d", "title": "2012 International Conference on Cyberworlds", "acronym": "cw", "groupId": "1000175", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNzFv4gM", "doi": "10.1109/CW.2012.36", "title": "Virtual Palpation for Medical Training in Cyberworlds", "normalizedTitle": "Virtual Palpation for Medical Training in Cyberworlds", "abstract": "In this paper, we introduce a new approach to virtual palpation for medical training in cyber worlds. We analyze palpation as a medical procedure. Then, we survey the existing virtual palpation projects, which use haptic devices, and propose a new image-driven approach to haptic palpation that can be easily ported to any web-enabled and collaborative environments. We also introduce variable haptic interaction point that allows us to implement multiple-point haptic interaction while using a single-point desktop haptic device. Lastly, we prove our hypothesis by implementing the proposed approach for abdominal palpation and validating it with medical practitioners. We also discuss the advantages of our method over other existing works in terms of flexibility, simplicity and scalability.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, we introduce a new approach to virtual palpation for medical training in cyber worlds. We analyze palpation as a medical procedure. Then, we survey the existing virtual palpation projects, which use haptic devices, and propose a new image-driven approach to haptic palpation that can be easily ported to any web-enabled and collaborative environments. We also introduce variable haptic interaction point that allows us to implement multiple-point haptic interaction while using a single-point desktop haptic device. Lastly, we prove our hypothesis by implementing the proposed approach for abdominal palpation and validating it with medical practitioners. We also discuss the advantages of our method over other existing works in terms of flexibility, simplicity and scalability.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, we introduce a new approach to virtual palpation for medical training in cyber worlds. We analyze palpation as a medical procedure. Then, we survey the existing virtual palpation projects, which use haptic devices, and propose a new image-driven approach to haptic palpation that can be easily ported to any web-enabled and collaborative environments. We also introduce variable haptic interaction point that allows us to implement multiple-point haptic interaction while using a single-point desktop haptic device. Lastly, we prove our hypothesis by implementing the proposed approach for abdominal palpation and validating it with medical practitioners. We also discuss the advantages of our method over other existing works in terms of flexibility, simplicity and scalability.", "fno": "4814a207", "keywords": [ "Haptic Interfaces", "Hip", "Force", "Fingers", "Visualization", "Biomedical Imaging", "Haptic Interaction Point", "Palpation", "Virtual Reality", "Haptic Device", "Function Based" ], "authors": [ { "affiliation": null, "fullName": "Shamima Yasmin", "givenName": "Shamima", "surname": "Yasmin", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Alexei Sourin", "givenName": "Alexei", "surname": "Sourin", "__typename": "ArticleAuthorType" } ], "idPrefix": "cw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-09-01T00:00:00", "pubType": "proceedings", "pages": "207-214", "year": "2012", "issn": null, "isbn": "978-1-4673-2736-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4814a199", "articleId": "12OmNzsrwkr", "__typename": "AdjacentArticleType" }, "next": { "fno": "4814a215", "articleId": "12OmNxAlA2N", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/achi/2009/3529/0/3529a193", "title": "Haptic Palpation for the Femoral Pulse in Virtual Interventional Radiology", "doi": null, "abstractUrl": "/proceedings-article/achi/2009/3529a193/12OmNAZOJU1", "parentPublication": { "id": "proceedings/achi/2009/3529/0", "title": "International Conference on Advances in Computer-Human Interaction", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2014/4261/0/4261a268", "title": "Breast Palpation Simulation with Haptic Feedback: Prototype and Initial Results", "doi": null, "abstractUrl": "/proceedings-article/svr/2014/4261a268/12OmNBhHt7R", "parentPublication": { "id": "proceedings/svr/2014/4261/0", "title": "2014 XVI Symposium on Virtual and Augmented Reality (SVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2017/2089/0/2089a202", "title": "Modeling Deformable Objects for Medical Training with Haptic Devices", "doi": null, "abstractUrl": "/proceedings-article/cw/2017/2089a202/12OmNCbU2U1", "parentPublication": { "id": "proceedings/cw/2017/2089/0", "title": "2017 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2010/9343/0/05643585", "title": "Haptic simulation of breast cancer palpation: A case study of haptic augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2010/05643585/12OmNwtn3ui", "parentPublication": { "id": "proceedings/ismar/2010/9343/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cisis/2016/0987/0/0987a382", "title": "Improvement of Mesh Free Deforming Analysis for Maxillofacial Palpation on a Virtual Training System", "doi": null, "abstractUrl": "/proceedings-article/cisis/2016/0987a382/12OmNxVDuQI", "parentPublication": { "id": "proceedings/cisis/2016/0987/0", "title": "2016 10th International Conference on Complex, Intelligent, and Software Intensive Systems (CISIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cbms/2011/1189/0/05999154", "title": "Ultrasound palpation by haptic elastography", "doi": null, "abstractUrl": "/proceedings-article/cbms/2011/05999154/12OmNy3Agyv", "parentPublication": { "id": "proceedings/cbms/2011/1189/0", "title": "2011 24th International Symposium on Computer-Based Medical Systems (CBMS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2012/01/tth2012010077", "title": "Rendering Virtual Tumors in Real Tissue Mock-Ups Using Haptic Augmented Reality", "doi": null, "abstractUrl": "/journal/th/2012/01/tth2012010077/13rRUwInvt1", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2012/02/tth2012020097", "title": "Abdominal Palpation Haptic Device for Colonoscopy Simulation Using Pneumatic Control", "doi": null, "abstractUrl": "/journal/th/2012/02/tth2012020097/13rRUwInvyJ", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040617", "title": "Haptic Palpation for Medical Simulation in Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040617/13rRUyfKIHI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNynsbxj", "title": "2014 XVI Symposium on Virtual and Augmented Reality (SVR)", "acronym": "svr", "groupId": "1800426", "volume": "0", "displayVolume": "0", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNzXnNpJ", "doi": "10.1109/SVR.2014.58", "title": "Systematically Reviewing Techniques and Devices Used in Palpation Simulation with Haptic Feedback", "normalizedTitle": "Systematically Reviewing Techniques and Devices Used in Palpation Simulation with Haptic Feedback", "abstract": "Palpation exam is a procedure where a health professional presses a specific region of a patient's body with fingers in order to detect the presence of features and abnormalities under the skin. There is a growing trend to develop simulators capable of adding improvements and greater realism to training this type of procedure. This article presents the results of a systematic review conducted in order to assess the state-of-art of the simulation of the palpation procedure. Papers focusing on techniques and tools of simulation palpation with hap tic feedback were analyzed. From the results obtained, it was concluded that despite being a new area, there are already many innovative papers that received positive reviews.", "abstracts": [ { "abstractType": "Regular", "content": "Palpation exam is a procedure where a health professional presses a specific region of a patient's body with fingers in order to detect the presence of features and abnormalities under the skin. There is a growing trend to develop simulators capable of adding improvements and greater realism to training this type of procedure. This article presents the results of a systematic review conducted in order to assess the state-of-art of the simulation of the palpation procedure. Papers focusing on techniques and tools of simulation palpation with hap tic feedback were analyzed. From the results obtained, it was concluded that despite being a new area, there are already many innovative papers that received positive reviews.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Palpation exam is a procedure where a health professional presses a specific region of a patient's body with fingers in order to detect the presence of features and abnormalities under the skin. There is a growing trend to develop simulators capable of adding improvements and greater realism to training this type of procedure. This article presents the results of a systematic review conducted in order to assess the state-of-art of the simulation of the palpation procedure. Papers focusing on techniques and tools of simulation palpation with hap tic feedback were analyzed. From the results obtained, it was concluded that despite being a new area, there are already many innovative papers that received positive reviews.", "fno": "4261a258", "keywords": [ "Haptic Interfaces", "Medical Diagnostic Computing", "Palpation Simulation", "Haptic Feedback", "Palpation Exam", "Health Professional", "Palpation Procedure", "Phantoms", "Haptic Interfaces", "Art", "Tumors", "Visualization", "IEEE Xplore", "Libraries", "Palpation", "Simulation", "Haptics", "Force Feedback", "Systematic Review" ], "authors": [ { "affiliation": "Lab. de Aplic. de Inf. em Saude (LApiS) Escola de Artes, Univ. de Sao Paulo, Sao Paulo, Brazil", "fullName": "Mateus de Lara Ribeiro", "givenName": "Mateus", "surname": "de Lara Ribeiro", "__typename": "ArticleAuthorType" }, { "affiliation": "Lab. de Aplic. de Inf. em Saude (LApiS) Escola de Artes, Univ. de Sao Paulo, Sao Paulo, Brazil", "fullName": "Fátima L.S. Nunes", "givenName": "Fátima L.S.", "surname": "Nunes", "__typename": "ArticleAuthorType" } ], "idPrefix": "svr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-05-01T00:00:00", "pubType": "proceedings", "pages": "258-267", "year": "2014", "issn": null, "isbn": "978-1-4799-4261-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4261a254", "articleId": "12OmNAolH8Q", "__typename": "AdjacentArticleType" }, "next": { "fno": "4261a268", "articleId": "12OmNBhHt7R", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/achi/2009/3529/0/3529a193", "title": "Haptic Palpation for the Femoral Pulse in Virtual Interventional Radiology", "doi": null, "abstractUrl": "/proceedings-article/achi/2009/3529a193/12OmNAZOJU1", "parentPublication": { "id": "proceedings/achi/2009/3529/0", "title": "International Conference on Advances in Computer-Human Interaction", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2014/4261/0/4261a268", "title": "Breast Palpation Simulation with Haptic Feedback: Prototype and Initial Results", "doi": null, "abstractUrl": "/proceedings-article/svr/2014/4261a268/12OmNBhHt7R", "parentPublication": { "id": "proceedings/svr/2014/4261/0", "title": "2014 XVI Symposium on Virtual and Augmented Reality (SVR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/waina/2015/1775/0/1775a044", "title": "Development of Virtual Palpation System for Dental Education", "doi": null, "abstractUrl": "/proceedings-article/waina/2015/1775a044/12OmNwMFMfZ", "parentPublication": { "id": "proceedings/waina/2015/1775/0", "title": "2015 IEEE 29th International Conference on Advanced Information Networking and Applications Workshops (WAINA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2010/9343/0/05643585", "title": "Haptic simulation of breast cancer palpation: A case study of haptic augmented reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2010/05643585/12OmNwtn3ui", "parentPublication": { "id": "proceedings/ismar/2010/9343/0", "title": "2010 IEEE International Symposium on Mixed and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cbms/2011/1189/0/05999154", "title": "Ultrasound palpation by haptic elastography", "doi": null, "abstractUrl": "/proceedings-article/cbms/2011/05999154/12OmNy3Agyv", "parentPublication": { "id": "proceedings/cbms/2011/1189/0", "title": "2011 24th International Symposium on Computer-Based Medical Systems (CBMS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2012/4814/0/4814a207", "title": "Virtual Palpation for Medical Training in Cyberworlds", "doi": null, "abstractUrl": "/proceedings-article/cw/2012/4814a207/12OmNzFv4gM", "parentPublication": { "id": "proceedings/cw/2012/4814/0", "title": "2012 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2012/02/tth2012020097", "title": "Abdominal Palpation Haptic Device for Colonoscopy Simulation Using Pneumatic Control", "doi": null, "abstractUrl": "/journal/th/2012/02/tth2012020097/13rRUwInvyJ", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2011/02/tth2011020078", "title": "A Comparative Study of Haptic Stiffness Identification by Veterinarians and Students", "doi": null, "abstractUrl": "/journal/th/2011/02/tth2011020078/13rRUxcbnHp", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040617", "title": "Haptic Palpation for Medical Simulation in Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040617/13rRUyfKIHI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/12/07272121", "title": "Precise Haptic Device Co-Location for Visuo-Haptic Augmented Reality", "doi": null, "abstractUrl": "/journal/tg/2015/12/07272121/13rRUygT7fe", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqMPfSj", "title": "2007 IEEE Symposium on 3D User Interfaces", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2007", "__typename": "ProceedingType" }, "article": { "id": "12OmNBubOUh", "doi": "10.1109/3DUI.2007.340769", "title": "Exploring 3D Interaction in Alternate Control-Display Space Mappings", "normalizedTitle": "Exploring 3D Interaction in Alternate Control-Display Space Mappings", "abstract": "The desire to have intuitive, seamless 3D interaction fuels research exploration into new approaches to 3D interaction. However, within these explorations we continue to rely on Brunelleschi's perspective for display and map the interactive control space directly into it without much thought on the effect that this default mapping has. In contrast, there are many possibilities for creating 3D interaction spaces, thus making it important to run user studies to examine these possibilities. Options in mapping the control space to the display space for 3D interaction have previously focused on the manipulation of control-display ratio or gain. In this paper, we present a conceptual framework that provides a more general control-display description that includes mappings for flip, rotation, skew, as well as scale (gain). We conduct a user study to explore 3D selection and manipulation tasks in three of these different mappings in comparison to the commonly used mapping (perspective mapping of control space to a perspective display). Our results show interesting differences between interactions and user preferences in these mappings and indicate that all may be considered viable alternatives. Together this framework and study open the door to further exploration of 3D interaction variations", "abstracts": [ { "abstractType": "Regular", "content": "The desire to have intuitive, seamless 3D interaction fuels research exploration into new approaches to 3D interaction. However, within these explorations we continue to rely on Brunelleschi's perspective for display and map the interactive control space directly into it without much thought on the effect that this default mapping has. In contrast, there are many possibilities for creating 3D interaction spaces, thus making it important to run user studies to examine these possibilities. Options in mapping the control space to the display space for 3D interaction have previously focused on the manipulation of control-display ratio or gain. In this paper, we present a conceptual framework that provides a more general control-display description that includes mappings for flip, rotation, skew, as well as scale (gain). We conduct a user study to explore 3D selection and manipulation tasks in three of these different mappings in comparison to the commonly used mapping (perspective mapping of control space to a perspective display). Our results show interesting differences between interactions and user preferences in these mappings and indicate that all may be considered viable alternatives. Together this framework and study open the door to further exploration of 3D interaction variations", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The desire to have intuitive, seamless 3D interaction fuels research exploration into new approaches to 3D interaction. However, within these explorations we continue to rely on Brunelleschi's perspective for display and map the interactive control space directly into it without much thought on the effect that this default mapping has. In contrast, there are many possibilities for creating 3D interaction spaces, thus making it important to run user studies to examine these possibilities. Options in mapping the control space to the display space for 3D interaction have previously focused on the manipulation of control-display ratio or gain. In this paper, we present a conceptual framework that provides a more general control-display description that includes mappings for flip, rotation, skew, as well as scale (gain). We conduct a user study to explore 3D selection and manipulation tasks in three of these different mappings in comparison to the commonly used mapping (perspective mapping of control space to a perspective display). Our results show interesting differences between interactions and user preferences in these mappings and indicate that all may be considered viable alternatives. Together this framework and study open the door to further exploration of 3D interaction variations", "fno": "04142840", "keywords": [ "Manipulation Task", "Control Display Space Mapping", "3 D Interaction Space", "Control Space", "Control Display Ratio", "Conceptual Framework", "3 D Selection" ], "authors": [ { "affiliation": "Dept. of Comput. Sci., Calgary Univ., Alta.", "fullName": "J. Keijser", "givenName": "J.", "surname": "Keijser", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Calgary Univ., Alta.", "fullName": "S. Carpendale", "givenName": "S.", "surname": "Carpendale", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Calgary Univ., Alta.", "fullName": "M. Hancock", "givenName": "M.", "surname": "Hancock", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Calgary Univ., Alta.", "fullName": "T. Isenberg", "givenName": "T.", "surname": "Isenberg", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2007-03-01T00:00:00", "pubType": "proceedings", "pages": "null", "year": "2007", "issn": null, "isbn": "1-4244-0907-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "04142839", "articleId": "12OmNxeM49A", "__typename": "AdjacentArticleType" }, "next": { "fno": "04142841", "articleId": "12OmNAP1YXh", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/apcip/2009/3699/2/3699b031", "title": "3D Multi-view Autostereoscopic Display and Its Key Technologie", "doi": null, "abstractUrl": "/proceedings-article/apcip/2009/3699b031/12OmNvAS4pe", "parentPublication": { "id": "proceedings/apcip/2009/3699/1", "title": "Information Processing, Asia-Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2006/2602/0/26020778", "title": "A Projection-Based Multi-view Time-multiplexed Autostereoscopic 3D Display System", "doi": null, "abstractUrl": "/proceedings-article/iv/2006/26020778/12OmNwDSdGX", "parentPublication": { "id": "proceedings/iv/2006/2602/0", "title": "Tenth International Conference on Information Visualisation (IV'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2007/0907/0/04142851", "title": "Character Interaction System with Autostereoscopic Display and Range Sensor", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142851/12OmNwc3wsr", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iih-msp/2009/3762/0/3762a174", "title": "Three-Dimensional Circular Camera System for the Three-Dimensional Display Based on the Reconstruction of Parallax Rays", "doi": null, "abstractUrl": "/proceedings-article/iih-msp/2009/3762a174/12OmNwlHT0Z", "parentPublication": { "id": "proceedings/iih-msp/2009/3762/0", "title": "Intelligent Information Hiding and Multimedia Signal Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2007/0907/0/04142839", "title": "An Exploration of Interaction-Display Offset in Surround Screen Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142839/12OmNxeM49A", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cisis/2008/3109/0/3109a763", "title": "PDA -- Remote Display Interaction Framework", "doi": null, "abstractUrl": "/proceedings-article/cisis/2008/3109a763/12OmNyL0TrV", "parentPublication": { "id": "proceedings/cisis/2008/3109/0", "title": "2008 International Conference on Complex, Intelligent and Software Intensive Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/c5/2005/2325/0/23250043", "title": "3D Data Transmitting and Displaying System with an Omni-Directional Display for Group Learning", "doi": null, "abstractUrl": "/proceedings-article/c5/2005/23250043/12OmNzZmZw6", "parentPublication": { "id": "proceedings/c5/2005/2325/0", "title": "Proceedings. Third International Conference on Creating, Connecting and Collaborating through Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/co/2005/08/r8046", "title": "Computer-Generated Holography as a Generic Display Technology", "doi": null, "abstractUrl": "/magazine/co/2005/08/r8046/13rRUB7a16k", "parentPublication": { "id": "mags/co", "title": "Computer", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2010/02/mpc2010020072", "title": "Human-Display Interaction Technology: Emerging Remote Interfaces for Pervasive Display Environments", "doi": null, "abstractUrl": "/magazine/pc/2010/02/mpc2010020072/13rRUxBa53l", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2009/02/mpc2009020014", "title": "Smart Phone Interaction with Registered Displays", "doi": null, "abstractUrl": "/magazine/pc/2009/02/mpc2009020014/13rRUyhaIm1", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAsTgX7", "title": "2010 IEEE Symposium on 3D User Interfaces (3DUI)", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNx3Zjcm", "doi": "10.1109/3DUI.2010.5444727", "title": "Evaluating natural interaction techniques in video games", "normalizedTitle": "Evaluating natural interaction techniques in video games", "abstract": "Despite the gaming industry's recent trend for using ¿natural¿ interaction techniques, which mimic real world actions with a high level of fidelity, it is not clear how natural interaction techniques affect the player experience. In order to obtain a better understanding, we designed and conducted a study using Mario Kart Wii, a commercial racing game for the Nintendo Wii. We chose this platform due to its seemingly balanced design of both natural and non-natural interaction techniques. Our empirical study of these techniques found that the non-natural interaction techniques significantly outperform their more natural counterparts. We offer three hypotheses to explain our finding and suggest them as important interaction design considerations.", "abstracts": [ { "abstractType": "Regular", "content": "Despite the gaming industry's recent trend for using ¿natural¿ interaction techniques, which mimic real world actions with a high level of fidelity, it is not clear how natural interaction techniques affect the player experience. In order to obtain a better understanding, we designed and conducted a study using Mario Kart Wii, a commercial racing game for the Nintendo Wii. We chose this platform due to its seemingly balanced design of both natural and non-natural interaction techniques. Our empirical study of these techniques found that the non-natural interaction techniques significantly outperform their more natural counterparts. We offer three hypotheses to explain our finding and suggest them as important interaction design considerations.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Despite the gaming industry's recent trend for using ¿natural¿ interaction techniques, which mimic real world actions with a high level of fidelity, it is not clear how natural interaction techniques affect the player experience. In order to obtain a better understanding, we designed and conducted a study using Mario Kart Wii, a commercial racing game for the Nintendo Wii. We chose this platform due to its seemingly balanced design of both natural and non-natural interaction techniques. Our empirical study of these techniques found that the non-natural interaction techniques significantly outperform their more natural counterparts. We offer three hypotheses to explain our finding and suggest them as important interaction design considerations.", "fno": "05444727", "keywords": [ "Nonnatural Interaction Techniques", "Natural Interaction Techniques", "Video Games", "Gaming Industry", "Commercial Racing Game", "Nintendo Wii" ], "authors": [ { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Ryan P McMahan", "givenName": "Ryan P", "surname": "McMahan", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Alexander Joel D Alon", "givenName": "Alexander Joel D", "surname": "Alon", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Shaimaa Lazem", "givenName": "Shaimaa", "surname": "Lazem", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Robert J Beaton", "givenName": "Robert J", "surname": "Beaton", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "David Machaj", "givenName": "David", "surname": "Machaj", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Michael Schaefer", "givenName": "Michael", "surname": "Schaefer", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Mara G Silva", "givenName": "Mara G", "surname": "Silva", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Anamary Leal", "givenName": "Anamary", "surname": "Leal", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Robert Hagan", "givenName": "Robert", "surname": "Hagan", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Virginia Tech, Blacksburg, VA, USA", "fullName": "Doug A Bowman", "givenName": "Doug A", "surname": "Bowman", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-03-01T00:00:00", "pubType": "proceedings", "pages": "11-14", "year": "2010", "issn": null, "isbn": "978-1-4244-6846-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05444737", "articleId": "12OmNrkBwIk", "__typename": "AdjacentArticleType" }, "next": { "fno": "05444728", "articleId": "12OmNzZmZtk", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ontoract/2008/3542/0/3542a091", "title": "Ontological Dimensions Applied to Natural Interaction", "doi": null, "abstractUrl": "/proceedings-article/ontoract/2008/3542a091/12OmNqGiu78", "parentPublication": { "id": "proceedings/ontoract/2008/3542/0", "title": "Ontologies in Interactive Systems, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2015/1727/0/07223386", "title": "Comparing the performance of natural, semi-natural, and non-natural locomotion techniques in virtual reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223386/12OmNx9nGM1", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/isip/2010/4261/0/4261a495", "title": "A Preliminary Study on Nonverbal Emotion Interaction of Virtual Characters in Cartoon Games", "doi": null, "abstractUrl": "/proceedings-article/isip/2010/4261a495/12OmNyv7m1C", "parentPublication": { "id": "proceedings/isip/2010/4261/0", "title": "2010 Third International Symposium on Information Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2008/05/mcg2008050010", "title": "Bringing VR and Spatial 3D Interaction to the Masses through Video Games", "doi": null, "abstractUrl": "/magazine/cg/2008/05/mcg2008050010/13rRUB6SpRY", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040626", "title": "Evaluating Display Fidelity and Interaction Fidelity in a Virtual Reality Game", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040626/13rRUwIF6dN", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2008/03/mpc2008030039", "title": "Hacking the Nintendo Wii Remote", "doi": null, "abstractUrl": "/magazine/pc/2008/03/mpc2008030039/13rRUx0xPsw", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/ic/2015/06/mic2015060060", "title": "Natural Interaction with Visualization Systems", "doi": null, "abstractUrl": "/magazine/ic/2015/06/mic2015060060/13rRUxCRFSG", "parentPublication": { "id": "mags/ic", "title": "IEEE Internet Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/06/v1224", "title": "Toward a Deeper Understanding of the Role of Interaction in Information Visualization", "doi": null, "abstractUrl": "/journal/tg/2007/06/v1224/13rRUxD9gXy", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/12/ttg2012122719", "title": "Interaction Support for Visual Comparison Inspired by Natural Behavior", "doi": null, "abstractUrl": "/journal/tg/2012/12/ttg2012122719/13rRUxZRbo0", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/isaiam/2021/3260/0/326000a001", "title": "Natural Interaction Technology in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/isaiam/2021/326000a001/1wiQXhFT4D6", "parentPublication": { "id": "proceedings/isaiam/2021/3260/0", "title": "2021 International Symposium on Artificial Intelligence and its Application on Media (ISAIAM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzuIjeQ", "title": "Near Field Communication, International Workshop on", "acronym": "nfc", "groupId": "1002809", "volume": "0", "displayVolume": "0", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNxXUhTH", "doi": "10.1109/NFC.2009.10", "title": "Exploring Expressive NFC-Based Mobile Phone Interaction with Large Dynamic Displays", "normalizedTitle": "Exploring Expressive NFC-Based Mobile Phone Interaction with Large Dynamic Displays", "abstract": "Inherent obstacles in current mobile applications are the limited input and output capabilities of mobile phones. In many ways, e.g. in terms of display capabilities and processing power, today’s mobile phones are quite versatile and provide opportunity for a multitude of new applications. Conversely, mobile phones will always be somewhat limited by certain form factors that are intrinsic to their compact design. For instance, small screen sizes make it difficult to visualize and manage applications that require a large amount of information on-screen for display and interaction. This paper reflects on our work accomplished when developing a new NFC interaction technique in which a mobile phone can be used as a direct input device for interaction with large dynamic displays. Using the (previously published) touch & interact interaction technique, the user touches the display with their mobile phone to perform an action. Beyond expressive interaction, an advantage of this approach is that large displays can provide spatial awareness while the mobile phone uses its capabilities to enhance the interaction (e.g. providing privacy and new interaction possibilities). We give technical details of our implementation and reflect on the most salient findings from three user studies, two of which elicit very promising results regarding selection performance and usability when compared with current, existing alternatives. Throughout the development of these prototypes, various interaction caveats have been identified, which are outlined in this paper, that promise to influence future NFC applications and interaction paradigms.", "abstracts": [ { "abstractType": "Regular", "content": "Inherent obstacles in current mobile applications are the limited input and output capabilities of mobile phones. In many ways, e.g. in terms of display capabilities and processing power, today’s mobile phones are quite versatile and provide opportunity for a multitude of new applications. Conversely, mobile phones will always be somewhat limited by certain form factors that are intrinsic to their compact design. For instance, small screen sizes make it difficult to visualize and manage applications that require a large amount of information on-screen for display and interaction. This paper reflects on our work accomplished when developing a new NFC interaction technique in which a mobile phone can be used as a direct input device for interaction with large dynamic displays. Using the (previously published) touch & interact interaction technique, the user touches the display with their mobile phone to perform an action. Beyond expressive interaction, an advantage of this approach is that large displays can provide spatial awareness while the mobile phone uses its capabilities to enhance the interaction (e.g. providing privacy and new interaction possibilities). We give technical details of our implementation and reflect on the most salient findings from three user studies, two of which elicit very promising results regarding selection performance and usability when compared with current, existing alternatives. Throughout the development of these prototypes, various interaction caveats have been identified, which are outlined in this paper, that promise to influence future NFC applications and interaction paradigms.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Inherent obstacles in current mobile applications are the limited input and output capabilities of mobile phones. In many ways, e.g. in terms of display capabilities and processing power, today’s mobile phones are quite versatile and provide opportunity for a multitude of new applications. Conversely, mobile phones will always be somewhat limited by certain form factors that are intrinsic to their compact design. For instance, small screen sizes make it difficult to visualize and manage applications that require a large amount of information on-screen for display and interaction. This paper reflects on our work accomplished when developing a new NFC interaction technique in which a mobile phone can be used as a direct input device for interaction with large dynamic displays. Using the (previously published) touch & interact interaction technique, the user touches the display with their mobile phone to perform an action. Beyond expressive interaction, an advantage of this approach is that large displays can provide spatial awareness while the mobile phone uses its capabilities to enhance the interaction (e.g. providing privacy and new interaction possibilities). We give technical details of our implementation and reflect on the most salient findings from three user studies, two of which elicit very promising results regarding selection performance and usability when compared with current, existing alternatives. Throughout the development of these prototypes, various interaction caveats have been identified, which are outlined in this paper, that promise to influence future NFC applications and interaction paradigms.", "fno": "3577a036", "keywords": [ "Mobile", "NFC", "Touch", "Display", "Interaction" ], "authors": [ { "affiliation": null, "fullName": "Robert Hardy", "givenName": "Robert", "surname": "Hardy", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Enrico Rukzio", "givenName": "Enrico", "surname": "Rukzio", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Matthias Wagner", "givenName": "Matthias", "surname": "Wagner", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Massimo Paolucci", "givenName": "Massimo", "surname": "Paolucci", "__typename": "ArticleAuthorType" } ], "idPrefix": "nfc", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-02-01T00:00:00", "pubType": "proceedings", "pages": "36-41", "year": "2009", "issn": null, "isbn": "978-0-7695-3577-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3577a030", "articleId": "12OmNz3bdKP", "__typename": "AdjacentArticleType" }, "next": { "fno": "3577a045", "articleId": "12OmNxGAL0y", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/nfc/2010/3998/0/3998a027", "title": "Mobile Prescription: An NFC-Based Proposal for AAL", "doi": null, "abstractUrl": "/proceedings-article/nfc/2010/3998a027/12OmNBV9IfU", "parentPublication": { "id": "proceedings/nfc/2010/3998/0", "title": "Near Field Communication, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nfc/2011/4327/0/4327a021", "title": "NFC Mobile Parlor Games Enabling Direct Player to Player Interaction", "doi": null, "abstractUrl": "/proceedings-article/nfc/2011/4327a021/12OmNvBrgHQ", "parentPublication": { "id": "proceedings/nfc/2011/4327/0", "title": "Near Field Communication, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nfc/2011/4327/0/4327a077", "title": "Practical Attacks on NFC Enabled Cell Phones", "doi": null, "abstractUrl": "/proceedings-article/nfc/2011/4327a077/12OmNwF0BM2", "parentPublication": { "id": "proceedings/nfc/2011/4327/0", "title": "Near Field Communication, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/uic-atc/2009/3737/0/3737a298", "title": "Toward a NFC Phone-Driven Context Awareness Smart Environment", "doi": null, "abstractUrl": "/proceedings-article/uic-atc/2009/3737a298/12OmNynJMUM", "parentPublication": { "id": "proceedings/uic-atc/2009/3737/0", "title": "Ubiquitous, Autonomic and Trusted Computing, Symposia and Workshops on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ares/2009/3564/0/3564a695", "title": "Vulnerability Analysis and Attacks on NFC-Enabled Mobile Phones", "doi": null, "abstractUrl": "/proceedings-article/ares/2009/3564a695/12OmNyuy9Li", "parentPublication": { "id": "proceedings/ares/2009/3564/0", "title": "2009 International Conference on Availability, Reliability and Security", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nfc/2009/3577/0/3577a030", "title": "Using NFC-Enabled Mobile Phones for Public Health in Developing Countries", "doi": null, "abstractUrl": "/proceedings-article/nfc/2009/3577a030/12OmNz3bdKP", "parentPublication": { "id": "proceedings/nfc/2009/3577/0", "title": "Near Field Communication, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nfc/2011/4327/0/4327a044", "title": "A Generic NFC-enabled Measurement System for Remote Monitoring and Control of Client-side Equipment", "doi": null, "abstractUrl": "/proceedings-article/nfc/2011/4327a044/12OmNzBwGIh", "parentPublication": { "id": "proceedings/nfc/2011/4327/0", "title": "Near Field Communication, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmb-gmr/2010/4084/0/4084a082", "title": "Deploying NFC Technology for Mobile Ticketing Services ", "doi": null, "abstractUrl": "/proceedings-article/icmb-gmr/2010/4084a082/12OmNzcxZqp", "parentPublication": { "id": "proceedings/icmb-gmr/2010/4084/0", "title": "Mobile Business / Global Mobility Roundtable, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nfc/2011/4327/0/4327a015", "title": "Touch to Play -- Exploring Touch-Based Mobile Interaction with Public Displays", "doi": null, "abstractUrl": "/proceedings-article/nfc/2011/4327a015/12OmNzn38XH", "parentPublication": { "id": "proceedings/nfc/2011/4327/0", "title": "Near Field Communication, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2009/02/mpc2009020014", "title": "Smart Phone Interaction with Registered Displays", "doi": null, "abstractUrl": "/magazine/pc/2009/02/mpc2009020014/13rRUyhaIm1", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqMPfSj", "title": "2007 IEEE Symposium on 3D User Interfaces", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2007", "__typename": "ProceedingType" }, "article": { "id": "12OmNxeM49A", "doi": "10.1109/3DUI.2007.340768", "title": "An Exploration of Interaction-Display Offset in Surround Screen Virtual Environments", "normalizedTitle": "An Exploration of Interaction-Display Offset in Surround Screen Virtual Environments", "abstract": "We present a study exploring the effect of positional offset between the user's interaction frame-of-reference (the physical location of input) and the display frame-of-reference (where graphical feedback appears) in a surround-screen virtual environment (SSVE). Our research hypothesis states that, in such an environment, task performance improves given an offset between the two frames-of-reference. In our experiment, users were asked to match a target color using a 3D color widget under three different display-interaction offset conditions: no offset (i.e., collocation), a three inch offset, and a two foot offset. Our results suggest that collocation of the display and interaction frames-of-reference may degrade accuracy in widget-based tasks and that collocation does not necessarily lead the user to spend more time on the task. In addition, these results contrast with previous studies performed with head-mounted display (HMD) platforms, which have demonstrated significant performance advantages for collocation and the \"direct manipulation\" of virtual objects. Moreover, a previous study with a different task performed in a projector-based VE has also demonstrated that collocation is not detrimental to user performance. Our conclusion is that the most effective positional offset is dependent upon the specific display hardware and VE task", "abstracts": [ { "abstractType": "Regular", "content": "We present a study exploring the effect of positional offset between the user's interaction frame-of-reference (the physical location of input) and the display frame-of-reference (where graphical feedback appears) in a surround-screen virtual environment (SSVE). Our research hypothesis states that, in such an environment, task performance improves given an offset between the two frames-of-reference. In our experiment, users were asked to match a target color using a 3D color widget under three different display-interaction offset conditions: no offset (i.e., collocation), a three inch offset, and a two foot offset. Our results suggest that collocation of the display and interaction frames-of-reference may degrade accuracy in widget-based tasks and that collocation does not necessarily lead the user to spend more time on the task. In addition, these results contrast with previous studies performed with head-mounted display (HMD) platforms, which have demonstrated significant performance advantages for collocation and the \"direct manipulation\" of virtual objects. Moreover, a previous study with a different task performed in a projector-based VE has also demonstrated that collocation is not detrimental to user performance. Our conclusion is that the most effective positional offset is dependent upon the specific display hardware and VE task", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We present a study exploring the effect of positional offset between the user's interaction frame-of-reference (the physical location of input) and the display frame-of-reference (where graphical feedback appears) in a surround-screen virtual environment (SSVE). Our research hypothesis states that, in such an environment, task performance improves given an offset between the two frames-of-reference. In our experiment, users were asked to match a target color using a 3D color widget under three different display-interaction offset conditions: no offset (i.e., collocation), a three inch offset, and a two foot offset. Our results suggest that collocation of the display and interaction frames-of-reference may degrade accuracy in widget-based tasks and that collocation does not necessarily lead the user to spend more time on the task. In addition, these results contrast with previous studies performed with head-mounted display (HMD) platforms, which have demonstrated significant performance advantages for collocation and the \"direct manipulation\" of virtual objects. Moreover, a previous study with a different task performed in a projector-based VE has also demonstrated that collocation is not detrimental to user performance. Our conclusion is that the most effective positional offset is dependent upon the specific display hardware and VE task", "fno": "04142839", "keywords": [ "Display Hardware", "Interaction Display Offset", "Surround Screen Virtual Environment", "User Interaction Frame Of Reference", "Display Frame Of Reference", "3 D Color Widget", "Display Interaction Offset Condition", "Head Mounted Display", "Virtual Object Manipulation", "Collocation" ], "authors": [ { "affiliation": "Dept. of Comput. Sci., Brown Univ., Providence, RI", "fullName": "D.K. Lemmerman", "givenName": "D.K.", "surname": "Lemmerman", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Brown Univ., Providence, RI", "fullName": "J.J. LaViola", "givenName": "J.J.", "surname": "LaViola", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2007-03-01T00:00:00", "pubType": "proceedings", "pages": "null", "year": "2007", "issn": null, "isbn": "1-4244-0907-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "04142838", "articleId": "12OmNwErpsL", "__typename": "AdjacentArticleType" }, "next": { "fno": "04142840", "articleId": "12OmNBubOUh", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/nbis/2011/4458/0/4458a456", "title": "A New Interface for Large Scale Tiled Display System Considering Scalability", "doi": null, "abstractUrl": "/proceedings-article/nbis/2011/4458a456/12OmNBU1jQj", "parentPublication": { "id": "proceedings/nbis/2011/4458/0", "title": "2011 14th International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2006/0224/0/02240281", "title": "A Foveal Inset for Large Display Environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2006/02240281/12OmNrkT7BT", "parentPublication": { "id": "proceedings/vr/2006/0224/0", "title": "IEEE Virtual Reality Conference (VR 2006)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/paccs/2009/3614/0/3614a135", "title": "Methods of Time Offset Estimation and Performance Evaluation of Synchronization for 3GPP LTE Downlink", "doi": null, "abstractUrl": "/proceedings-article/paccs/2009/3614a135/12OmNvRU0q9", "parentPublication": { "id": "proceedings/paccs/2009/3614/0", "title": "2009 Pacific-Asia Conference on Circuits, Communications and Systems (PACCS 2009)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/apwcs/2010/4003/0/4003a066", "title": "A Solution of LED Large Screen Display Based on Wireless Communication", "doi": null, "abstractUrl": "/proceedings-article/apwcs/2010/4003a066/12OmNvq5jzx", "parentPublication": { "id": "proceedings/apwcs/2010/4003/0", "title": "Wearable Computing Systems, Asia-Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wcse/2009/3570/2/3570b239", "title": "An Ontology Based Prototype for Geocoding Offset Addresses", "doi": null, "abstractUrl": "/proceedings-article/wcse/2009/3570b239/12OmNx5GU9M", "parentPublication": { "id": null, "title": null, "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vts/2009/3598/0/3598a185", "title": "A Scalable, Digital BIST Circuit for Measurement and Compensation of Static Phase Offset", "doi": null, "abstractUrl": "/proceedings-article/vts/2009/3598a185/12OmNxdm4yA", "parentPublication": { "id": "proceedings/vts/2009/3598/0", "title": "2009 27th IEEE VLSI Test Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icinis/2010/4249/0/4249a217", "title": "ARM9-Based Control System for LED Large Screen Display", "doi": null, "abstractUrl": "/proceedings-article/icinis/2010/4249a217/12OmNxuo0kI", "parentPublication": { "id": "proceedings/icinis/2010/4249/0", "title": "Intelligent Networks and Intelligent Systems, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2007/0905/0/04161055", "title": "Effects of Interaction-Display Offset on User Performance in Surround Screen Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2007/04161055/12OmNxymo84", "parentPublication": { "id": "proceedings/vr/2007/0905/0", "title": "2007 IEEE Virtual Reality Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1993/1363/0/00380765", "title": "Sparcchair: A one hundred million pixel display", "doi": null, "abstractUrl": "/proceedings-article/vrais/1993/00380765/12OmNyrqzDT", "parentPublication": { "id": "proceedings/vrais/1993/1363/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2010/02/mpc2010020072", "title": "Human-Display Interaction Technology: Emerging Remote Interfaces for Pervasive Display Environments", "doi": null, "abstractUrl": "/magazine/pc/2010/02/mpc2010020072/13rRUxBa53l", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyfdOJ3", "title": "2008 International Conference on Complex, Intelligent and Software Intensive Systems", "acronym": "cisis", "groupId": "1002005", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNyL0TrV", "doi": "10.1109/CISIS.2008.130", "title": "PDA -- Remote Display Interaction Framework", "normalizedTitle": "PDA -- Remote Display Interaction Framework", "abstract": "This paper discusses the design of a PDA-driven remote display control system which was designed for a pervasive computing scenario. Our solution enables a PDA to perform as an interaction device (such as keyboard or mouse) to operate applications which are using remote displays. Our system consists of two components: one devoted to pair users and remote display according to their relative position, and one to provide the PDA with appropriate graphical interfaces and ad-hoc communication protocols. Besides a case study on the use of remote displays during a course within our department, this paper discusses system highlights, components, and experiments as for standard multimedia and office applications.", "abstracts": [ { "abstractType": "Regular", "content": "This paper discusses the design of a PDA-driven remote display control system which was designed for a pervasive computing scenario. Our solution enables a PDA to perform as an interaction device (such as keyboard or mouse) to operate applications which are using remote displays. Our system consists of two components: one devoted to pair users and remote display according to their relative position, and one to provide the PDA with appropriate graphical interfaces and ad-hoc communication protocols. Besides a case study on the use of remote displays during a course within our department, this paper discusses system highlights, components, and experiments as for standard multimedia and office applications.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper discusses the design of a PDA-driven remote display control system which was designed for a pervasive computing scenario. Our solution enables a PDA to perform as an interaction device (such as keyboard or mouse) to operate applications which are using remote displays. Our system consists of two components: one devoted to pair users and remote display according to their relative position, and one to provide the PDA with appropriate graphical interfaces and ad-hoc communication protocols. Besides a case study on the use of remote displays during a course within our department, this paper discusses system highlights, components, and experiments as for standard multimedia and office applications.", "fno": "3109a763", "keywords": [ "Human Environment Interface", "Remote Display", "Pervasive Systems" ], "authors": [ { "affiliation": null, "fullName": "Alessandro Genco", "givenName": "Alessandro", "surname": "Genco", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Salvatore Sorce", "givenName": "Salvatore", "surname": "Sorce", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Rosario Messineo", "givenName": "Rosario", "surname": "Messineo", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Paolo Raccuglia", "givenName": "Paolo", "surname": "Raccuglia", "__typename": "ArticleAuthorType" } ], "idPrefix": "cisis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-03-01T00:00:00", "pubType": "proceedings", "pages": "763-768", "year": "2008", "issn": null, "isbn": "978-0-7695-3109-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3109a757", "articleId": "12OmNyQ7G0N", "__typename": "AdjacentArticleType" }, "next": { "fno": "3109a769", "articleId": "12OmNC3Xhkg", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3pgcic/2011/4531/0/4531a299", "title": "Experiment on Multi-video Transmission with Multipoint Tiled Display Wall", "doi": null, "abstractUrl": "/proceedings-article/3pgcic/2011/4531a299/12OmNAXPye0", "parentPublication": { "id": "proceedings/3pgcic/2011/4531/0", "title": "P2P, Parallel, Grid, Cloud, and Internet Computing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nbis/2011/4458/0/4458a444", "title": "Proposal of High Realistic Sensation System Using the Large Scale Tiled Display Environment", "doi": null, "abstractUrl": "/proceedings-article/nbis/2011/4458a444/12OmNB6D71g", "parentPublication": { "id": "proceedings/nbis/2011/4458/0", "title": "2011 14th International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpads/2005/2281/2/228120167", "title": "Interactive Controller for 3D Contents with Omni-directional Display", "doi": null, "abstractUrl": "/proceedings-article/icpads/2005/228120167/12OmNwwd2Py", "parentPublication": { "id": "proceedings/icpads/2005/2281/1", "title": "Parallel and Distributed Systems, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cisis/2009/3575/0/3575a789", "title": "Embedding the Remote Application Control: Visual API for PDA Programming", "doi": null, "abstractUrl": "/proceedings-article/cisis/2009/3575a789/12OmNxGj9Zh", "parentPublication": { "id": "proceedings/cisis/2009/3575/0", "title": "2009 International Conference on Complex, Intelligent and Software Intensive Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3pgcic/2010/4237/0/4237a421", "title": "An Experiment on Multipoint Tele-immersive Communication with Tiled Display Wall", "doi": null, "abstractUrl": "/proceedings-article/3pgcic/2010/4237a421/12OmNy2ah07", "parentPublication": { "id": "proceedings/3pgcic/2010/4237/0", "title": "P2P, Parallel, Grid, Cloud, and Internet Computing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2002/1492/0/14920081", "title": "A Testbed for Precise Registration, Natural Occlusion and Interaction in an Augmented Environment Using a Head-Mounted Projective Display (HMPD)", "doi": null, "abstractUrl": "/proceedings-article/vr/2002/14920081/12OmNylboC4", "parentPublication": { "id": "proceedings/vr/2002/1492/0", "title": "Proceedings IEEE Virtual Reality 2002", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tc/1976/12/01674600", "title": "Trends in Graphic Display Design", "doi": null, "abstractUrl": "/journal/tc/1976/12/01674600/13rRUwh80FL", "parentPublication": { "id": "trans/tc", "title": "IEEE Transactions on Computers", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2010/02/mpc2010020072", "title": "Human-Display Interaction Technology: Emerging Remote Interfaces for Pervasive Display Environments", "doi": null, "abstractUrl": "/magazine/pc/2010/02/mpc2010020072/13rRUxBa53l", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/co/2011/08/mco2011080046", "title": "Remote Display Solutions for Mobile Cloud Computing", "doi": null, "abstractUrl": "/magazine/co/2011/08/mco2011080046/13rRUy0HYNe", "parentPublication": { "id": "mags/co", "title": "Computer", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNCwUmAj", "title": "2008 Second International Symposium on Universal Communication", "acronym": "isuc", "groupId": "1002556", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNz2kqm0", "doi": "10.1109/ISUC.2008.73", "title": "Inertial Force Display - Concept and Implementation", "normalizedTitle": "Inertial Force Display - Concept and Implementation", "abstract": "This paper discusses an approach to implementation of a force display that presents sensation of inertial force. Concept of object-oriented haptic interface is proposed, an idea of information transmission through haptic interaction is discussed, a prototype mechanism and its control system is designed, and implementation of the prototype system is described. Through experiment on presenting simple object models, it was confirmed that the prototype system works as expected.", "abstracts": [ { "abstractType": "Regular", "content": "This paper discusses an approach to implementation of a force display that presents sensation of inertial force. Concept of object-oriented haptic interface is proposed, an idea of information transmission through haptic interaction is discussed, a prototype mechanism and its control system is designed, and implementation of the prototype system is described. Through experiment on presenting simple object models, it was confirmed that the prototype system works as expected.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper discusses an approach to implementation of a force display that presents sensation of inertial force. Concept of object-oriented haptic interface is proposed, an idea of information transmission through haptic interaction is discussed, a prototype mechanism and its control system is designed, and implementation of the prototype system is described. Through experiment on presenting simple object models, it was confirmed that the prototype system works as expected.", "fno": "3433a281", "keywords": [ "Force Display", "Object Oriented Device", "Haptic Interaction" ], "authors": [ { "affiliation": null, "fullName": "Koichi Hirota", "givenName": "Koichi", "surname": "Hirota", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yuichiro Sekiguchi", "givenName": "Yuichiro", "surname": "Sekiguchi", "__typename": "ArticleAuthorType" } ], "idPrefix": "isuc", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-12-01T00:00:00", "pubType": "proceedings", "pages": "281-284", "year": "2008", "issn": null, "isbn": "978-0-7695-3433-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3433a274", "articleId": "12OmNwO5LYZ", "__typename": "AdjacentArticleType" }, "next": { "fno": "3433a285", "articleId": "12OmNBtCCIi", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/haptics/2008/2005/0/04479973", "title": "Interactive representation of virtual object in hand-held box by finger-worn haptic display", "doi": null, "abstractUrl": "/proceedings-article/haptics/2008/04479973/12OmNAHmOtG", "parentPublication": { "id": "proceedings/haptics/2008/2005/0", "title": "IEEE Haptics Symposium 2008", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2011/4602/0/4602a259", "title": "Feasibility of a Tele-rehabilitation System with Corrective-Force", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2011/4602a259/12OmNCdk2ET", "parentPublication": { "id": "proceedings/icvrv/2011/4602/0", "title": "2011 International Conference on Virtual Reality and Visualization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2006/0224/0/02240269", "title": "Proactive Desk II: Development of a New Multi-object Haptic Display Using a Linear Induction Motor", "doi": null, "abstractUrl": "/proceedings-article/vr/2006/02240269/12OmNvT2p0z", "parentPublication": { "id": "proceedings/vr/2006/0224/0", "title": "IEEE Virtual Reality Conference (VR 2006)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1997/7843/0/78430188", "title": "Haptic display of visual images", "doi": null, "abstractUrl": "/proceedings-article/vrais/1997/78430188/12OmNwIYZyx", "parentPublication": { "id": "proceedings/vrais/1997/7843/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/whc/2007/2738/0/04145247", "title": "Development of Fingertip Type Non-grounding Force Feedback Display", "doi": null, "abstractUrl": "/proceedings-article/whc/2007/04145247/12OmNx5Yv8J", "parentPublication": { "id": "proceedings/whc/2007/2738/0", "title": "2007 2nd Joint EuroHaptics Conference and Symposium on Haptic Interfaces for Virtual Environments and Teleoperator Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cisis/2010/3967/0/3967b158", "title": "A Haptic Parameter Exploration Method for Force Feedback Devices", "doi": null, "abstractUrl": "/proceedings-article/cisis/2010/3967b158/12OmNx8wThN", "parentPublication": { "id": "proceedings/cisis/2010/3967/0", "title": "2010 International Conference on Complex, Intelligent and Software Intensive Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2001/0948/0/09480123", "title": "HapticGEAR: The Development of a Wearable Force Display System for Immersive Projection Displays", "doi": null, "abstractUrl": "/proceedings-article/vr/2001/09480123/12OmNxGja0K", "parentPublication": { "id": "proceedings/vr/2001/0948/0", "title": "Virtual Reality Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/th/2017/03/07556272", "title": "Vibrotactile Compliance Feedback for Tangential Force Interaction", "doi": null, "abstractUrl": "/journal/th/2017/03/07556272/13rRUwcS1D9", "parentPublication": { "id": "trans/th", "title": "IEEE Transactions on Haptics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/1995/05/mcg1995050022", "title": "Providing Force Feedback in Virtual Environments", "doi": null, "abstractUrl": "/magazine/cg/1995/05/mcg1995050022/13rRUwfqpFZ", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2022/5365/0/536500a905", "title": "Haptics in VR Using Origami-Augmented Drones", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2022/536500a905/1J7WrPcWIVO", "parentPublication": { "id": "proceedings/ismar-adjunct/2022/5365/0", "title": "2022 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1KBdPyLx2W4", "title": "2022 International Conference on Wearables, Sports and Lifestyle Management (WSLM)", "acronym": "wslm", "groupId": "1848892", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1KBdRVssvBe", "doi": "10.1109/WSLM54683.2022.00012", "title": "Design of Radar Display System Based on Human-computer Interaction Interface", "normalizedTitle": "Design of Radar Display System Based on Human-computer Interaction Interface", "abstract": "Because the man-made natural operation control methods such as voice or gesture are prone to ambiguity or more complicated codec, in the practical exploration, scientific researchers have proposed a human-computer interaction interface, which is mainly combined with the operation method of human-computer cooperation to realize man-machine Interactive. This interface design has a positive effect on the radar display system and helps to present more effective application modules during system operation. This paper mainly analyzes how to use MATLAB software in radar system simulation. Among them, GUI human-computer interaction interface tool realizes simulation analysis of signal processing and power calculation in the overall design of radar. The final experimental results prove that this simulation software interface is convenient and reliable, and it can build a more excellent design platform for the work of the overall radar designer.", "abstracts": [ { "abstractType": "Regular", "content": "Because the man-made natural operation control methods such as voice or gesture are prone to ambiguity or more complicated codec, in the practical exploration, scientific researchers have proposed a human-computer interaction interface, which is mainly combined with the operation method of human-computer cooperation to realize man-machine Interactive. This interface design has a positive effect on the radar display system and helps to present more effective application modules during system operation. This paper mainly analyzes how to use MATLAB software in radar system simulation. Among them, GUI human-computer interaction interface tool realizes simulation analysis of signal processing and power calculation in the overall design of radar. The final experimental results prove that this simulation software interface is convenient and reliable, and it can build a more excellent design platform for the work of the overall radar designer.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Because the man-made natural operation control methods such as voice or gesture are prone to ambiguity or more complicated codec, in the practical exploration, scientific researchers have proposed a human-computer interaction interface, which is mainly combined with the operation method of human-computer cooperation to realize man-machine Interactive. This interface design has a positive effect on the radar display system and helps to present more effective application modules during system operation. This paper mainly analyzes how to use MATLAB software in radar system simulation. Among them, GUI human-computer interaction interface tool realizes simulation analysis of signal processing and power calculation in the overall design of radar. The final experimental results prove that this simulation software interface is convenient and reliable, and it can build a more excellent design platform for the work of the overall radar designer.", "fno": "081900a038", "keywords": [ "Computer Simulation", "Design Engineering", "Graphical User Interfaces", "Human Computer Interaction", "Man Machine Systems", "Matlab", "Radar Computing", "Radar Signal Processing", "GUI", "Human Computer Cooperation", "Human Computer Interaction Interface", "Interface Design", "Man Machine Interaction", "Man Made Natural Operation Control Methods", "MATLAB Software", "Radar Display System Design", "Radar System Simulation", "Simulation Software Interface", "System Operation", "Human Computer Interaction", "Training", "Analytical Models", "Wearable Computers", "Systems Operation", "Display Systems", "Software", "Human Computer Interaction Interface", "Software", "Radar Display System", "Tracking Accuracy Organization Of The Text" ], "authors": [ { "affiliation": "Xi'an Technological University,Xi'an,China", "fullName": "Chenhao Li", "givenName": "Chenhao", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "Xi'an Technological University,Xi'an,China", "fullName": "Bailin Liu", "givenName": "Bailin", "surname": "Liu", "__typename": "ArticleAuthorType" } ], "idPrefix": "wslm", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-01-01T00:00:00", "pubType": "proceedings", "pages": "38-42", "year": "2022", "issn": null, "isbn": "978-1-6654-0819-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "081900a033", "articleId": "1KBdQ0q0ezK", "__typename": "AdjacentArticleType" }, "next": { "fno": "081900a043", "articleId": "1KBdRNpJi3S", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icesssymposia/2008/3288/0/3288a028", "title": "A Radar Display Terminal Based on a Partial-Screen-Updating Method", "doi": null, "abstractUrl": "/proceedings-article/icesssymposia/2008/3288a028/12OmNAkWvh5", "parentPublication": { "id": "proceedings/icesssymposia/2008/3288/0", "title": "Embedded Software and Systems, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iis/1997/8218/0/82180090", "title": "Towards an adaptive man-machine interface for virtual environments", "doi": null, "abstractUrl": "/proceedings-article/iis/1997/82180090/12OmNAsk4Gp", "parentPublication": { "id": "proceedings/iis/1997/8218/0", "title": "Intelligent Information Systems, IASTED International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccae/2009/3569/0/3569a230", "title": "Human-Computer Interaction: Process and Principles of Human-Computer Interface Design", "doi": null, "abstractUrl": "/proceedings-article/iccae/2009/3569a230/12OmNrJAdRj", "parentPublication": { "id": "proceedings/iccae/2009/3569/0", "title": "2009 International Conference on Computer and Automation Engineering. ICCAE 2009", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mmcs/1997/7819/0/00609773", "title": "Automated lip synchronisation for human-computer interaction and special effect animation", "doi": null, "abstractUrl": "/proceedings-article/mmcs/1997/00609773/12OmNvlg8pF", "parentPublication": { "id": "proceedings/mmcs/1997/7819/0", "title": "Proceedings of IEEE International Conference on Multimedia Computing and Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/etcs/2009/3557/2/3557c437", "title": "Human-Computer Interaction Design in Product Design", "doi": null, "abstractUrl": "/proceedings-article/etcs/2009/3557c437/12OmNwMXnsg", "parentPublication": { "id": null, "title": null, "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/taai/2011/4601/0/4601a245", "title": "A Wearable Dichoptic Display System", "doi": null, "abstractUrl": "/proceedings-article/taai/2011/4601a245/12OmNzsrwlK", "parentPublication": { "id": "proceedings/taai/2011/4601/0", "title": "2011 International Conference on Technologies and Applications of Artificial Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2010/02/mpc2010020072", "title": "Human-Display Interaction Technology: Emerging Remote Interfaces for Pervasive Display Environments", "doi": null, "abstractUrl": "/magazine/pc/2010/02/mpc2010020072/13rRUxBa53l", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icedcs/2022/5541/0/554100a152", "title": "Application of Human-Computer Interaction Technology in Mobile Interface Design for Digital Media", "doi": null, "abstractUrl": "/proceedings-article/icedcs/2022/554100a152/1JC1t8SWn5u", "parentPublication": { "id": "proceedings/icedcs/2022/5541/0", "title": "2022 International Conference on Electronics and Devices, Computational Science (ICEDCS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/percom-workshops/2020/4716/0/09156151", "title": "Capturing Human Pose Using mmWave Radar", "doi": null, "abstractUrl": "/proceedings-article/percom-workshops/2020/09156151/1m1jFy1YhBm", "parentPublication": { "id": "proceedings/percom-workshops/2020/4716/0", "title": "2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ichci/2020/2316/0/231600a005", "title": "Human-Computer Interface Design Based on Design Psychology", "doi": null, "abstractUrl": "/proceedings-article/ichci/2020/231600a005/1tuAaot3gv6", "parentPublication": { "id": "proceedings/ichci/2020/2316/0", "title": "2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1ap5wvyUHKM", "title": "2017 International Conference on Virtual Reality and Visualization (ICVRV)", "acronym": "icvrv", "groupId": "1800579", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "1ap5wIzl6i4", "doi": "10.1109/ICVRV.2017.00063", "title": "Efficiency Group Interaction Between Participants and Large Display", "normalizedTitle": "Efficiency Group Interaction Between Participants and Large Display", "abstract": "In recent years there has been an increasing interest in human-computer interaction research in large display systems. We contribute with a design of efficient group interaction technique and application to a large display by using optical technology to track the markers on users' hands. Participants were asked to execute predefined interaction actions based on the contents of the application screen. With the help of well-designed interaction processes and algorithms, our system can receive response signals of multiple users and then trigger corresponding actions in real-time. Moreover, we designed a performance experiment to evaluate the recognition and response results from practical application. The experiment results demonstrate that, although integrating group interaction into large display system, this framework can effectively support user's participation in large display.", "abstracts": [ { "abstractType": "Regular", "content": "In recent years there has been an increasing interest in human-computer interaction research in large display systems. We contribute with a design of efficient group interaction technique and application to a large display by using optical technology to track the markers on users' hands. Participants were asked to execute predefined interaction actions based on the contents of the application screen. With the help of well-designed interaction processes and algorithms, our system can receive response signals of multiple users and then trigger corresponding actions in real-time. Moreover, we designed a performance experiment to evaluate the recognition and response results from practical application. The experiment results demonstrate that, although integrating group interaction into large display system, this framework can effectively support user's participation in large display.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In recent years there has been an increasing interest in human-computer interaction research in large display systems. We contribute with a design of efficient group interaction technique and application to a large display by using optical technology to track the markers on users' hands. Participants were asked to execute predefined interaction actions based on the contents of the application screen. With the help of well-designed interaction processes and algorithms, our system can receive response signals of multiple users and then trigger corresponding actions in real-time. Moreover, we designed a performance experiment to evaluate the recognition and response results from practical application. The experiment results demonstrate that, although integrating group interaction into large display system, this framework can effectively support user's participation in large display.", "fno": "263600a274", "keywords": [ "Display Instrumentation", "Groupware", "Human Computer Interaction", "Large Display Systems", "Predefined Interaction Actions", "Optical Technology", "Efficient Group Interaction Technique", "Human Computer Interaction Research", "Participants", "Response Signals", "Application Screen", "Display Systems", "Real Time Systems", "Rendering Computer Graphics", "Brightness", "Virtual Reality", "Visualization", "Cultural Differences", "Human Computer Interaction", "Large Display", "Multiple Users Application" ], "authors": [ { "affiliation": null, "fullName": "Hao Jiang", "givenName": "Hao", "surname": "Jiang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Chang Gao", "givenName": "Chang", "surname": "Gao", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Tianlu Mao", "givenName": "Tianlu", "surname": "Mao", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Hui Li", "givenName": "Hui", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zhaoqi Wang", "givenName": "Zhaoqi", "surname": "Wang", "__typename": "ArticleAuthorType" } ], "idPrefix": "icvrv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-10-01T00:00:00", "pubType": "proceedings", "pages": "274-275", "year": "2017", "issn": "2375-141X", "isbn": "978-1-5386-2636-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "263600a272", "articleId": "1ap5zBoUE48", "__typename": "AdjacentArticleType" }, "next": { "fno": "263600a276", "articleId": "1ap5DbTADlu", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dui/2007/0907/0/04142840", "title": "Exploring 3D Interaction in Alternate Control-Display Space Mappings", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142840/12OmNBubOUh", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3pgcic/2010/4237/0/4237a446", "title": "Tiled Display System for Improved Communication Efficiency", "doi": null, "abstractUrl": "/proceedings-article/3pgcic/2010/4237a446/12OmNqFrGMu", "parentPublication": { "id": "proceedings/3pgcic/2010/4237/0", "title": "P2P, Parallel, Grid, Cloud, and Internet Computing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1993/1363/0/00380766", "title": "An interactive graphics display architecture", "doi": null, "abstractUrl": "/proceedings-article/vrais/1993/00380766/12OmNvTjZUe", "parentPublication": { "id": "proceedings/vrais/1993/1363/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2007/0907/0/04142839", "title": "An Exploration of Interaction-Display Offset in Surround Screen Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142839/12OmNxeM49A", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cisis/2008/3109/0/3109a763", "title": "PDA -- Remote Display Interaction Framework", "doi": null, "abstractUrl": "/proceedings-article/cisis/2008/3109a763/12OmNyL0TrV", "parentPublication": { "id": "proceedings/cisis/2008/3109/0", "title": "2008 International Conference on Complex, Intelligent and Software Intensive Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ace/2010/4058/0/4058a316", "title": "High End Scientific Visualization with Scalable Display System", "doi": null, "abstractUrl": "/proceedings-article/ace/2010/4058a316/12OmNzmclE4", "parentPublication": { "id": "proceedings/ace/2010/4058/0", "title": "Advances in Computer Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2016/0836/0/07504718", "title": "Groupnect: Integrating group interaction into large display system", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504718/12OmNzxPTKw", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040626", "title": "Evaluating Display Fidelity and Interaction Fidelity in a Virtual Reality Game", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040626/13rRUwIF6dN", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/pc/2010/02/mpc2010020072", "title": "Human-Display Interaction Technology: Emerging Remote Interfaces for Pervasive Display Environments", "doi": null, "abstractUrl": "/magazine/pc/2010/02/mpc2010020072/13rRUxBa53l", "parentPublication": { "id": "mags/pc", "title": "IEEE Pervasive Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wslm/2022/0819/0/081900a038", "title": "Design of Radar Display System Based on Human-computer Interaction Interface", "doi": null, "abstractUrl": "/proceedings-article/wslm/2022/081900a038/1KBdRVssvBe", "parentPublication": { "id": "proceedings/wslm/2022/0819/0", "title": "2022 International Conference on Wearables, Sports and Lifestyle Management (WSLM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvjgWM6", "title": "2011 IEEE Symposium on 3D User Interfaces (3DUI)", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNB1wkHF", "doi": "10.1109/3DUI.2011.5759219", "title": "Rapid and accurate 3D selection by progressive refinement", "normalizedTitle": "Rapid and accurate 3D selection by progressive refinement", "abstract": "Issues such as hand and tracker jitter negatively affect user performance with the ray-casting selection technique in 3D environments. This makes it difficult for users to perform tasks that require them to select objects that have a small visible area, since small targets require high levels of precision. We introduce an approach to address this issue that uses progressive refinement of the set of selectable objects to reduce the required precision of the task. We present a design space of progressive refinement techniques and an exemplar technique called Sphere-casting refined by QUAD-menu (SQUAD). We explore the tradeoffs between progressive refinement and immediate selection techniques in an evaluation comparing SQUAD to ray-casting. Both an analytical evaluation based on a distal pointing model and an empirical evaluation demonstrate that progressive refinement selection can be better than immediate selection. SQUAD was much more accurate than ray-casting, and SQUAD was faster than ray-casting with small targets and less cluttered environments.", "abstracts": [ { "abstractType": "Regular", "content": "Issues such as hand and tracker jitter negatively affect user performance with the ray-casting selection technique in 3D environments. This makes it difficult for users to perform tasks that require them to select objects that have a small visible area, since small targets require high levels of precision. We introduce an approach to address this issue that uses progressive refinement of the set of selectable objects to reduce the required precision of the task. We present a design space of progressive refinement techniques and an exemplar technique called Sphere-casting refined by QUAD-menu (SQUAD). We explore the tradeoffs between progressive refinement and immediate selection techniques in an evaluation comparing SQUAD to ray-casting. Both an analytical evaluation based on a distal pointing model and an empirical evaluation demonstrate that progressive refinement selection can be better than immediate selection. SQUAD was much more accurate than ray-casting, and SQUAD was faster than ray-casting with small targets and less cluttered environments.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Issues such as hand and tracker jitter negatively affect user performance with the ray-casting selection technique in 3D environments. This makes it difficult for users to perform tasks that require them to select objects that have a small visible area, since small targets require high levels of precision. We introduce an approach to address this issue that uses progressive refinement of the set of selectable objects to reduce the required precision of the task. We present a design space of progressive refinement techniques and an exemplar technique called Sphere-casting refined by QUAD-menu (SQUAD). We explore the tradeoffs between progressive refinement and immediate selection techniques in an evaluation comparing SQUAD to ray-casting. Both an analytical evaluation based on a distal pointing model and an empirical evaluation demonstrate that progressive refinement selection can be better than immediate selection. SQUAD was much more accurate than ray-casting, and SQUAD was faster than ray-casting with small targets and less cluttered environments.", "fno": "05759219", "keywords": [], "authors": [ { "affiliation": "Dept. of Computer Science and Center for Human-Computer Interaction, Virginia Tech, USA", "fullName": "Regis Kopper", "givenName": "Regis", "surname": "Kopper", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Computer Science and Center for Human-Computer Interaction, Virginia Tech, USA", "fullName": "Felipe Bacim", "givenName": "Felipe", "surname": "Bacim", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Computer Science and Center for Human-Computer Interaction, Virginia Tech, USA", "fullName": "Doug A. Bowman", "givenName": "Doug A.", "surname": "Bowman", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-03-01T00:00:00", "pubType": "proceedings", "pages": "67-74", "year": "2011", "issn": null, "isbn": "978-1-4577-0063-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05759218", "articleId": "12OmNCxbXKZ", "__typename": "AdjacentArticleType" }, "next": { "fno": "05759220", "articleId": "12OmNvCRgje", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2013/4795/0/06549388", "title": "Towards a disambiguation canvas", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549388/12OmNBNM8RP", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2011/4548/0/4548a093", "title": "Accurate Volume Rendering of Unstructured Hexahedral Meshes", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2011/4548a093/12OmNCcbE5T", "parentPublication": { "id": "proceedings/sibgrapi/2011/4548/0", "title": "2011 24th SIBGRAPI Conference on Graphics, Patterns and Images", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pg/2001/1227/0/12270285", "title": "Adaptive Progressive Vertex Tracing in Distributed Environments", "doi": null, "abstractUrl": "/proceedings-article/pg/2001/12270285/12OmNqyUUv4", "parentPublication": { "id": "proceedings/pg/2001/1227/0", "title": "Computer Graphics and Applications, Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2017/6716/0/07893359", "title": "PRECIOUS! Out-of-reach selection using iterative refinement in VR", "doi": null, "abstractUrl": "/proceedings-article/3dui/2017/07893359/12OmNzFv4da", "parentPublication": { "id": "proceedings/3dui/2017/6716/0", "title": "2017 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/prs/1997/8265/0/82650039", "title": "Parallel progressive rendering of animation sequences at interactive rates on distributed-memory machines", "doi": null, "abstractUrl": "/proceedings-article/prs/1997/82650039/12OmNzdoMqC", "parentPublication": { "id": "proceedings/prs/1997/8265/0", "title": "Parallel Rendering Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/prs/1993/4920/0/00586089", "title": "Progressive refinement radiosity on ring-connected multicomputers", "doi": null, "abstractUrl": "/proceedings-article/prs/1993/00586089/12OmNzmclCE", "parentPublication": { "id": "proceedings/prs/1993/4920/0", "title": "Proceedings of 1993 IEEE Parallel Rendering Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wcse/2009/3570/4/3570d170", "title": "A Progressive Refinement Approach for JAVA Regression Test Selection", "doi": null, "abstractUrl": "/proceedings-article/wcse/2009/3570d170/12OmNzxPTIm", "parentPublication": { "id": "proceedings/wcse/2009/3570/4", "title": "2009 WRI World Congress on Software Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2022/8402/0/840200a582", "title": "Multi-Touch Smartphone-Based Progressive Refinement VR Selection", "doi": null, "abstractUrl": "/proceedings-article/vrw/2022/840200a582/1CJcBfmyX5K", "parentPublication": { "id": "proceedings/vrw/2022/8402/0", "title": "2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2022/8402/0/840200a304", "title": "IMPReSS: Improved Multi-Touch Progressive Refinement Selection Strategy", "doi": null, "abstractUrl": "/proceedings-article/vrw/2022/840200a304/1CJetSxfyi4", "parentPublication": { "id": "proceedings/vrw/2022/8402/0", "title": "2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2022/9062/0/09956111", "title": "Focus and Adjust: Progressive Refinement Network for Human Object Interaction Detection", "doi": null, "abstractUrl": "/proceedings-article/icpr/2022/09956111/1IHpckHoe3u", "parentPublication": { "id": "proceedings/icpr/2022/9062/0", "title": "2022 26th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyKJiaX", "title": "Multimedia Software Engineering, International Symposium on", "acronym": "ismse", "groupId": "1000484", "volume": "0", "displayVolume": "0", "year": "2004", "__typename": "ProceedingType" }, "article": { "id": "12OmNBC8Atb", "doi": "10.1109/MMSE.2004.32", "title": "Automatic Selection and Combination of Descriptors for Effective 3D Similarity Search", "normalizedTitle": "Automatic Selection and Combination of Descriptors for Effective 3D Similarity Search", "abstract": "We focus on improving the effectiveness of similarity search in 3D object repositories from a system-oriented perspective. Motivated by an effectiveness evaluation of several individual 3D retrieval methods, we research a selection heuristic, called purity, for choosing retrieval methods based on query-dependent characteristics. We show that the purity selection method significantly improves the search effectiveness compared to the best single methods. We then show that retrieval effectiveness can be further boosted by considering combinations of multiple retrieval methods to perform the search. We propose to use a dynamically weighted combination of feature vectors based on the purity concept, and we experimentally show that the search effectiveness of our combined methods by far exceeds the effectiveness of our best implemented single method.", "abstracts": [ { "abstractType": "Regular", "content": "We focus on improving the effectiveness of similarity search in 3D object repositories from a system-oriented perspective. Motivated by an effectiveness evaluation of several individual 3D retrieval methods, we research a selection heuristic, called purity, for choosing retrieval methods based on query-dependent characteristics. We show that the purity selection method significantly improves the search effectiveness compared to the best single methods. We then show that retrieval effectiveness can be further boosted by considering combinations of multiple retrieval methods to perform the search. We propose to use a dynamically weighted combination of feature vectors based on the purity concept, and we experimentally show that the search effectiveness of our combined methods by far exceeds the effectiveness of our best implemented single method.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We focus on improving the effectiveness of similarity search in 3D object repositories from a system-oriented perspective. Motivated by an effectiveness evaluation of several individual 3D retrieval methods, we research a selection heuristic, called purity, for choosing retrieval methods based on query-dependent characteristics. We show that the purity selection method significantly improves the search effectiveness compared to the best single methods. We then show that retrieval effectiveness can be further boosted by considering combinations of multiple retrieval methods to perform the search. We propose to use a dynamically weighted combination of feature vectors based on the purity concept, and we experimentally show that the search effectiveness of our combined methods by far exceeds the effectiveness of our best implemented single method.", "fno": "22170514", "keywords": [ "3 D Objects", "Information Retrieval", "Query By Content", "Effectiveness", "Feature Selection" ], "authors": [ { "affiliation": "University of Konstanz", "fullName": "Benjamin Bustos", "givenName": "Benjamin", "surname": "Bustos", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Konstanz", "fullName": "Daniel Keim", "givenName": "Daniel", "surname": "Keim", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Konstanz", "fullName": "Dietmar Saupe", "givenName": "Dietmar", "surname": "Saupe", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Konstanz", "fullName": "Tobias Schreck", "givenName": "Tobias", "surname": "Schreck", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Konstanz", "fullName": "Dejan Vranić", "givenName": "Dejan", "surname": "Vranić", "__typename": "ArticleAuthorType" } ], "idPrefix": "ismse", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2004-12-01T00:00:00", "pubType": "proceedings", "pages": "514-521", "year": "2004", "issn": null, "isbn": "0-7695-2217-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "22170507", "articleId": "12OmNBr4evc", "__typename": "AdjacentArticleType" }, "next": { "fno": "22170522", "articleId": "12OmNCyBXj2", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2015/8391/0/8391b044", "title": "Adaptive Hashing for Fast Similarity Search", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391b044/12OmNC1Y5jM", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sisap/2008/3101/0/3101a105", "title": "A Dynamic Pivot Selection Technique for Similarity Search", "doi": null, "abstractUrl": "/proceedings-article/sisap/2008/3101a105/12OmNCbU2V9", "parentPublication": { "id": "proceedings/sisap/2008/3101/0", "title": "2008 First International Workshop on Similarity Search and Applications (SISAP '08)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2004/8603/2/01394465", "title": "Using entropy impurity for improved 3D object similarity search", "doi": null, "abstractUrl": "/proceedings-article/icme/2004/01394465/12OmNvSbBHg", "parentPublication": { "id": "proceedings/icme/2004/8603/2", "title": "2004 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2006/2746/0/274600799", "title": "Fast Similarity Search for High-Dimensional Dataset", "doi": null, "abstractUrl": "/proceedings-article/ism/2006/274600799/12OmNvUsoqa", "parentPublication": { "id": "proceedings/ism/2006/2746/0", "title": "Eighth IEEE International Symposium on Multimedia (ISM'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2007/1016/0/04285063", "title": "Are Concept Detector Lexicons Effective for Video Search?", "doi": null, "abstractUrl": "/proceedings-article/icme/2007/04285063/12OmNwDACbI", "parentPublication": { "id": "proceedings/icme/2007/1016/0", "title": "2007 International Conference on Multimedia & Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icde/2009/3545/0/3545a517", "title": "Effective XML Keyword Search with Relevance Oriented Ranking", "doi": null, "abstractUrl": "/proceedings-article/icde/2009/3545a517/12OmNwvDQvV", "parentPublication": { "id": "proceedings/icde/2009/3545/0", "title": "2009 IEEE 25th International Conference on Data Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmcs/1999/0253/1/02539512", "title": "Relevance Feedback and Category Search in Image Databases", "doi": null, "abstractUrl": "/proceedings-article/icmcs/1999/02539512/12OmNzYwc4J", "parentPublication": { "id": "proceedings/icmcs/1999/0253/1", "title": "Multimedia Computing and Systems, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tk/2004/01/k0028", "title": "Personalized Web Search For Improving Retrieval Effectiveness", "doi": null, "abstractUrl": "/journal/tk/2004/01/k0028/13rRUIM2VHj", "parentPublication": { "id": "trans/tk", "title": "IEEE Transactions on Knowledge & Data Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tk/2010/08/ttk2010081077", "title": "Towards an Effective XML Keyword Search", "doi": null, "abstractUrl": "/journal/tk/2010/08/ttk2010081077/13rRUxcsYMm", "parentPublication": { "id": "trans/tk", "title": "IEEE Transactions on Knowledge & Data Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tb/2016/05/07321808", "title": "Similarity Search of Flexible 3D Molecules Combining Local and Global Shape Descriptors", "doi": null, "abstractUrl": "/journal/tb/2016/05/07321808/13rRUy2YLRy", "parentPublication": { "id": "trans/tb", "title": "IEEE/ACM Transactions on Computational Biology and Bioinformatics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwCJOXn", "title": "3D User Interfaces (3DUI'06)", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2006", "__typename": "ProceedingType" }, "article": { "id": "12OmNBZpH8h", "doi": "10.1109/VR.2006.140", "title": "Using the Non-Dominant Hand for Selection in 3D", "normalizedTitle": "Using the Non-Dominant Hand for Selection in 3D", "abstract": "Although 3D virtual environments are designed to provide the user with an intuitive interface to view or manipulate highly complex data, current solutions are still not ideal. In order to make the interaction as natural as possible, metaphors are used to allow the users to apply their everyday knowledge in the generated environment. In literature, a lot of experiments can be found, describing new or improved metaphors. In our former work, we presented the ?Object In Hand? metaphor [4], which addresses some problems regarding the access of objects and menus in a 3D world. Although the metaphor turned out to be very promising, the solution shifted the problem towards a selection problem. From the insights of our previous work, we believe the non-dominant hand can play a role in solving this problem. In this paper we formally compare three well-known selection metaphors and we will check their suitability to be carried out with the non-dominant hand in order to seamlessly integrate the most suitable selection metaphor within the ?Object In Hand? metaphor.", "abstracts": [ { "abstractType": "Regular", "content": "Although 3D virtual environments are designed to provide the user with an intuitive interface to view or manipulate highly complex data, current solutions are still not ideal. In order to make the interaction as natural as possible, metaphors are used to allow the users to apply their everyday knowledge in the generated environment. In literature, a lot of experiments can be found, describing new or improved metaphors. In our former work, we presented the ?Object In Hand? metaphor [4], which addresses some problems regarding the access of objects and menus in a 3D world. Although the metaphor turned out to be very promising, the solution shifted the problem towards a selection problem. From the insights of our previous work, we believe the non-dominant hand can play a role in solving this problem. In this paper we formally compare three well-known selection metaphors and we will check their suitability to be carried out with the non-dominant hand in order to seamlessly integrate the most suitable selection metaphor within the ?Object In Hand? metaphor.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Although 3D virtual environments are designed to provide the user with an intuitive interface to view or manipulate highly complex data, current solutions are still not ideal. In order to make the interaction as natural as possible, metaphors are used to allow the users to apply their everyday knowledge in the generated environment. In literature, a lot of experiments can be found, describing new or improved metaphors. In our former work, we presented the ?Object In Hand? metaphor [4], which addresses some problems regarding the access of objects and menus in a 3D world. Although the metaphor turned out to be very promising, the solution shifted the problem towards a selection problem. From the insights of our previous work, we believe the non-dominant hand can play a role in solving this problem. In this paper we formally compare three well-known selection metaphors and we will check their suitability to be carried out with the non-dominant hand in order to seamlessly integrate the most suitable selection metaphor within the ?Object In Hand? metaphor.", "fno": "02250053", "keywords": [ "3 D Virtual Environments", "Selection Metaphors", "User S Experiment" ], "authors": [ { "affiliation": "Hasselt University, Expertise centre for Digital Media and transnationale Universiteit Limburg, Belgium", "fullName": "Joan De Boeck", "givenName": "Joan", "surname": "De Boeck", "__typename": "ArticleAuthorType" }, { "affiliation": "Hasselt University, Expertise centre for Digital Media and transnationale Universiteit Limburg, Belgium", "fullName": "Tom De Weyer", "givenName": "Tom", "surname": "De Weyer", "__typename": "ArticleAuthorType" }, { "affiliation": "Hasselt University, Expertise centre for Digital Media and transnationale Universiteit Limburg, Belgium", "fullName": "Chris Raymaekers", "givenName": "Chris", "surname": "Raymaekers", "__typename": "ArticleAuthorType" }, { "affiliation": "Hasselt University, Expertise centre for Digital Media and transnationale Universiteit Limburg, Belgium", "fullName": "Karin Coninx", "givenName": "Karin", "surname": "Coninx", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2006-03-01T00:00:00", "pubType": "proceedings", "pages": "53-58", "year": "2006", "issn": null, "isbn": "1-4244-0225-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "02250045", "articleId": "12OmNylKB1L", "__typename": "AdjacentArticleType" }, "next": { "fno": "02250062", "articleId": "12OmNwHQBbr", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dui/2007/0907/0/04142849", "title": "Balloon Selection: A Multi-Finger Technique for Accurate Low-Fatigue 3D Selection", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142849/12OmNBp52GJ", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444783", "title": "GPU implementation of 3D object selection by conic volume techniques in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444783/12OmNBp52w1", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2017/6647/0/07892295", "title": "Gauntlet: Travel technique for immersive environments using non-dominant hand", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892295/12OmNCxbXAC", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-infovis/2000/0804/0/08040155", "title": "Metaphor-Aware 3D Navigation", "doi": null, "abstractUrl": "/proceedings-article/ieee-infovis/2000/08040155/12OmNqIhFWx", "parentPublication": { "id": "proceedings/ieee-infovis/2000/0804/0", "title": "Information Visualization, IEEE Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iiai-aai/2015/9957/0/07373999", "title": "Difference in Bi-articular Muscle during Postural Fixation between Dominant and Non-dominant Arms", "doi": null, "abstractUrl": "/proceedings-article/iiai-aai/2015/07373999/12OmNwEJ12A", "parentPublication": { "id": "proceedings/iiai-aai/2015/9957/0", "title": "2015 IIAI 4th International Congress on Advanced Applied Informatics (IIAI-AAI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2000/0743/0/07430379", "title": "Mapping Information onto 3D Virtual Worlds", "doi": null, "abstractUrl": "/proceedings-article/iv/2000/07430379/12OmNxRWI8j", "parentPublication": { "id": "proceedings/iv/2000/0743/0", "title": "2000 IEEE Conference on Information Visualization. An International Conference on Computer Visualization and Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2008/2047/0/04476604", "title": "Poster: Sliding Viewport for Head Mounted Displays in Interactive Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2008/04476604/12OmNzdoMAW", "parentPublication": { "id": "proceedings/3dui/2008/2047/0", "title": "2008 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040634", "title": "Dense and Dynamic 3D Selection for Game-Based Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040634/13rRUyYSWsS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sbgames/2018/9605/0/960500a127", "title": "Evaluation of Selection Techniques on a Mobile Augmented Reality Game", "doi": null, "abstractUrl": "/proceedings-article/sbgames/2018/960500a127/17D45VVho5D", "parentPublication": { "id": "proceedings/sbgames/2018/9605/0", "title": "2018 17th Brazilian Symposium on Computer Games and Digital Entertainment (SBGames)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2023/05/10049645", "title": "GestureSurface: VR Sketching through Assembling Scaffold Surface with Non-Dominant Hand", "doi": null, "abstractUrl": "/journal/tg/2023/05/10049645/1KYoyLX55fy", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqMPfSj", "title": "2007 IEEE Symposium on 3D User Interfaces", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2007", "__typename": "ProceedingType" }, "article": { "id": "12OmNBp52GJ", "doi": "10.1109/3DUI.2007.340778", "title": "Balloon Selection: A Multi-Finger Technique for Accurate Low-Fatigue 3D Selection", "normalizedTitle": "Balloon Selection: A Multi-Finger Technique for Accurate Low-Fatigue 3D Selection", "abstract": "Balloon selection is a 3D interaction technique that is modeled after the real world metaphor of manipulating a helium balloon attached to a string. Balloon selection allows for precise 3D selection in the volume above a tabletop surface by using multiple fingers on a multi-touch-sensitive surface. The 3DOF selection tasks is decomposed in part into a 2DOF positioning task performed by one finger on the tabletop in an absolute 2D Cartesian coordinate system and a 1DOF positioning task performed by another finger on the tabletop in a relative 2D polar coordinate system. We have evaluated balloon selection in a formal user study that compared it to two well-known interaction techniques for selecting a static 3D target: a 3DOF tracked wand and keyboard cursor keys. We found that balloon selection was significantly faster than using cursor keys and had a significantly lower error rate than the wand. The lower error rate appeared to result from the user's hands being supported by the tabletop surface, resulting in significantly reduced hand tremor and arm fatigue.", "abstracts": [ { "abstractType": "Regular", "content": "Balloon selection is a 3D interaction technique that is modeled after the real world metaphor of manipulating a helium balloon attached to a string. Balloon selection allows for precise 3D selection in the volume above a tabletop surface by using multiple fingers on a multi-touch-sensitive surface. The 3DOF selection tasks is decomposed in part into a 2DOF positioning task performed by one finger on the tabletop in an absolute 2D Cartesian coordinate system and a 1DOF positioning task performed by another finger on the tabletop in a relative 2D polar coordinate system. We have evaluated balloon selection in a formal user study that compared it to two well-known interaction techniques for selecting a static 3D target: a 3DOF tracked wand and keyboard cursor keys. We found that balloon selection was significantly faster than using cursor keys and had a significantly lower error rate than the wand. The lower error rate appeared to result from the user's hands being supported by the tabletop surface, resulting in significantly reduced hand tremor and arm fatigue.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Balloon selection is a 3D interaction technique that is modeled after the real world metaphor of manipulating a helium balloon attached to a string. Balloon selection allows for precise 3D selection in the volume above a tabletop surface by using multiple fingers on a multi-touch-sensitive surface. The 3DOF selection tasks is decomposed in part into a 2DOF positioning task performed by one finger on the tabletop in an absolute 2D Cartesian coordinate system and a 1DOF positioning task performed by another finger on the tabletop in a relative 2D polar coordinate system. We have evaluated balloon selection in a formal user study that compared it to two well-known interaction techniques for selecting a static 3D target: a 3DOF tracked wand and keyboard cursor keys. We found that balloon selection was significantly faster than using cursor keys and had a significantly lower error rate than the wand. The lower error rate appeared to result from the user's hands being supported by the tabletop surface, resulting in significantly reduced hand tremor and arm fatigue.", "fno": "04142849", "keywords": [ "Keyboard Cursor Key", "Multifinger Technique", "Accurate Low Fatigue 3 D Selection", "3 D Interaction", "Tabletop Surface", "Multiple Finger", "Multitouch Sensitive Surface", "3 DOF Selection", "2 DOF Positioning Task", "2 D Cartesian Coordinate System", "1 DOF Positioning Task", "2 D Polar Coordinate System", "Balloon Selection Evaluation", "Static 3 D Target", "3 DOF Tracked Wand" ], "authors": [ { "affiliation": "Dept. of Comput. Sci., Columbia Univ., New York, NY", "fullName": "H. Benko", "givenName": "H.", "surname": "Benko", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Comput. Sci., Columbia Univ., New York, NY", "fullName": "S. Feiner", "givenName": "S.", "surname": "Feiner", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2007-03-01T00:00:00", "pubType": "proceedings", "pages": "null", "year": "2007", "issn": null, "isbn": "1-4244-0907-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "04142848", "articleId": "12OmNyrIatw", "__typename": "AdjacentArticleType" }, "next": { "fno": "04142850", "articleId": "12OmNvrMUfz", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dui/2006/0225/0/02250053", "title": "Using the Non-Dominant Hand for Selection in 3D", "doi": null, "abstractUrl": "/proceedings-article/3dui/2006/02250053/12OmNBZpH8h", "parentPublication": { "id": "proceedings/3dui/2006/0225/0", "title": "3D User Interfaces (3DUI'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444783", "title": "GPU implementation of 3D object selection by conic volume techniques in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444783/12OmNBp52w1", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sccc/2010/4400/0/4400a218", "title": "Selection Techniques for Dense and Occluded Virtual 3D Environments, Supported by Depth Feedback: Double, Bound and Depth Bubble Cursors", "doi": null, "abstractUrl": "/proceedings-article/sccc/2010/4400a218/12OmNrAdsv3", "parentPublication": { "id": "proceedings/sccc/2010/4400/0", "title": "2010 XXIX International Conference of the Chilean Computer Science Society", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iita/2009/3859/3/3859c339", "title": "Interested Sample Point Pre-Selection Based Dense Terrain Reconstruction for Autonomous Navigation", "doi": null, "abstractUrl": "/proceedings-article/iita/2009/3859c339/12OmNs5rkX9", "parentPublication": { "id": "proceedings/iita/2009/3859/3", "title": "2009 Third International Symposium on Intelligent Information Technology Application", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dicta/2011/4588/0/4588a689", "title": "Cooperative Relay Selection Based UEP Scheme for 3D Video Transmission over Rayleigh Fading Channel", "doi": null, "abstractUrl": "/proceedings-article/dicta/2011/4588a689/12OmNwErpTZ", "parentPublication": { "id": "proceedings/dicta/2011/4588/0", "title": "2011 International Conference on Digital Image Computing: Techniques and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2006/0225/0/02250103", "title": "Towards a General Model for Selection in Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2006/02250103/12OmNwF0BWC", "parentPublication": { "id": "proceedings/3dui/2006/0225/0", "title": "3D User Interfaces (3DUI'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2007/0907/0/04142854", "title": "Exploring the Effects of Environment Density and Target Visibility on Object Selection in 3D Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142854/12OmNwK7o7o", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2006/06/i0983", "title": "Two-View Multibody Structure-and-Motion with Outliers through Model Selection", "doi": null, "abstractUrl": "/journal/tp/2006/06/i0983/13rRUwkfB0l", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/12/ttg2012122245", "title": "Efficient Structure-Aware Selection Techniques for 3D Point Cloud Visualizations with 2DOF Input", "doi": null, "abstractUrl": "/journal/tg/2012/12/ttg2012122245/13rRUxASupy", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040634", "title": "Dense and Dynamic 3D Selection for Game-Based Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040634/13rRUyYSWsS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBC8AAD", "title": "2010 IEEE Virtual Reality Conference (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNBp52w1", "doi": "10.1109/VR.2010.5444783", "title": "GPU implementation of 3D object selection by conic volume techniques in virtual environments", "normalizedTitle": "GPU implementation of 3D object selection by conic volume techniques in virtual environments", "abstract": "In this paper we present a GPU implementation to accurately select 3D objects based on their silhouettes by a pointing device with six degrees of freedom (6DOF) in a virtual environment (VE). We adapt a 2D picking metaphor to 3D selection in VE's by changing the projection and view matrices according to the position and orientation of a 6DOF pointing device and rendering a conic selection volume to an off-screen pixel buffer. This method works for triangulated as well as volume rendered objects, no explicit geometric representation is required.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper we present a GPU implementation to accurately select 3D objects based on their silhouettes by a pointing device with six degrees of freedom (6DOF) in a virtual environment (VE). We adapt a 2D picking metaphor to 3D selection in VE's by changing the projection and view matrices according to the position and orientation of a 6DOF pointing device and rendering a conic selection volume to an off-screen pixel buffer. This method works for triangulated as well as volume rendered objects, no explicit geometric representation is required.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper we present a GPU implementation to accurately select 3D objects based on their silhouettes by a pointing device with six degrees of freedom (6DOF) in a virtual environment (VE). We adapt a 2D picking metaphor to 3D selection in VE's by changing the projection and view matrices according to the position and orientation of a 6DOF pointing device and rendering a conic selection volume to an off-screen pixel buffer. This method works for triangulated as well as volume rendered objects, no explicit geometric representation is required.", "fno": "05444783", "keywords": [ "Six Degrees Of Freedom", "GPU Implementation", "3 D Object Selection", "Conic Selection Volume Techniques", "Virtual Environments" ], "authors": [ { "affiliation": "Virtual Reality Group JARA, RWTH Aachen Univ., Aachen, Germany", "fullName": "Tobias Rick", "givenName": "Tobias", "surname": "Rick", "__typename": "ArticleAuthorType" }, { "affiliation": "Virtual Reality Group JARA, RWTH Aachen Univ., Aachen, Germany", "fullName": "Anette von Kapri", "givenName": "Anette", "surname": "von Kapri", "__typename": "ArticleAuthorType" }, { "affiliation": "Virtual Reality Group JARA, RWTH Aachen Univ., Aachen, Germany", "fullName": "Torsten Kuhlen", "givenName": "Torsten", "surname": "Kuhlen", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-03-01T00:00:00", "pubType": "proceedings", "pages": "243-246", "year": "2010", "issn": null, "isbn": "978-1-4244-6237-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05444782", "articleId": "12OmNBO3JYm", "__typename": "AdjacentArticleType" }, "next": { "fno": "05444780", "articleId": "12OmNyQph3i", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/sibgrapi/2011/4548/0/4548a056", "title": "Geodesic Conic Subdivision Curves on Surfaces", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2011/4548a056/12OmNBU1jVb", "parentPublication": { "id": "proceedings/sibgrapi/2011/4548/0", "title": "2011 24th SIBGRAPI Conference on Graphics, Patterns and Images", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2003/2030/0/20300038", "title": "Acceleration Techniques for GPU-based Volume Rendering", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2003/20300038/12OmNC2xhD8", "parentPublication": { "id": "proceedings/ieee-vis/2003/2030/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iros/1995/7108/3/71083291", "title": "Geometric design of hexapod with integrated limb mechanism of leg and arm", "doi": null, "abstractUrl": "/proceedings-article/iros/1995/71083291/12OmNvDZERN", "parentPublication": { "id": "proceedings/iros/1995/7108/3", "title": "Intelligent Robots and Systems, IEEE/RSJ International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/1995/7062/0/70620056", "title": "Motion planning for computer animation and virtual reality applications", "doi": null, "abstractUrl": "/proceedings-article/ca/1995/70620056/12OmNxEBzmx", "parentPublication": { "id": "proceedings/ca/1995/7062/0", "title": "Computer Animation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2003/2030/0/01250384", "title": "Acceleration techniques for GPU-based volume rendering", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2003/01250384/12OmNxeM48Y", "parentPublication": { "id": "proceedings/ieee-vis/2003/2030/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzDvSnO", "title": "2010 XXIX International Conference of the Chilean Computer Science Society", "acronym": "sccc", "groupId": "1001489", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNrAdsv3", "doi": "10.1109/SCCC.2010.51", "title": "Selection Techniques for Dense and Occluded Virtual 3D Environments, Supported by Depth Feedback: Double, Bound and Depth Bubble Cursors", "normalizedTitle": "Selection Techniques for Dense and Occluded Virtual 3D Environments, Supported by Depth Feedback: Double, Bound and Depth Bubble Cursors", "abstract": "In 3D Graphic Environments the selection of an object can be hampered by other objects. This occurs when the object becomes occluded by others. The probability of this happening is high if there are many objects in the scene. Over the years, this problem has been addressed by developing various techniques of selection, in order to provide a good mechanism to manipulate objects in a 3D virtual environment. These techniques have been evolving as they solved the problems of their predecessors. Although the third dimension is the distinguishing characteristic of these environments, the best among them does not take depth feedback as an important factor. Therefore, this document addresses this deficiency through the creation of three techniques based on that feedback.", "abstracts": [ { "abstractType": "Regular", "content": "In 3D Graphic Environments the selection of an object can be hampered by other objects. This occurs when the object becomes occluded by others. The probability of this happening is high if there are many objects in the scene. Over the years, this problem has been addressed by developing various techniques of selection, in order to provide a good mechanism to manipulate objects in a 3D virtual environment. These techniques have been evolving as they solved the problems of their predecessors. Although the third dimension is the distinguishing characteristic of these environments, the best among them does not take depth feedback as an important factor. Therefore, this document addresses this deficiency through the creation of three techniques based on that feedback.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In 3D Graphic Environments the selection of an object can be hampered by other objects. This occurs when the object becomes occluded by others. The probability of this happening is high if there are many objects in the scene. Over the years, this problem has been addressed by developing various techniques of selection, in order to provide a good mechanism to manipulate objects in a 3D virtual environment. These techniques have been evolving as they solved the problems of their predecessors. Although the third dimension is the distinguishing characteristic of these environments, the best among them does not take depth feedback as an important factor. Therefore, this document addresses this deficiency through the creation of three techniques based on that feedback.", "fno": "4400a218", "keywords": [ "Virtual Reality", "3 D Interaction", "Selection", "Depth Feedback" ], "authors": [ { "affiliation": null, "fullName": "Daniel Alejandro Winkler Rosa", "givenName": "Daniel Alejandro Winkler", "surname": "Rosa", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Hubert Hoffmann Nagel", "givenName": "Hubert Hoffmann", "surname": "Nagel", "__typename": "ArticleAuthorType" } ], "idPrefix": "sccc", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-11-01T00:00:00", "pubType": "proceedings", "pages": "218-225", "year": "2010", "issn": "1522-4902", "isbn": "978-0-7695-4400-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4400a212", "articleId": "12OmNylbovO", "__typename": "AdjacentArticleType" }, "next": { "fno": "4400a226", "articleId": "12OmNyuyacQ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2006/0224/0/02240261", "title": "A Tangible User Interface System for CAVE Applicat", "doi": null, "abstractUrl": "/proceedings-article/vr/2006/02240261/12OmNrGKest", "parentPublication": { "id": "proceedings/vr/2006/0224/0", "title": "IEEE Virtual Reality Conference (VR 2006)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2006/0225/0/02250103", "title": "Towards a General Model for Selection in Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2006/02250103/12OmNwF0BWC", "parentPublication": { "id": "proceedings/3dui/2006/0225/0", "title": "3D User Interfaces (3DUI'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrais/1997/7843/0/78430188", "title": "Haptic display of visual images", "doi": null, "abstractUrl": "/proceedings-article/vrais/1997/78430188/12OmNwIYZyx", "parentPublication": { "id": "proceedings/vrais/1997/7843/0", "title": "Virtual Reality Annual International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2007/0907/0/04142854", "title": "Exploring the Effects of Environment Density and Target Visibility on Object Selection in 3D Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142854/12OmNwK7o7o", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cisis/2011/4373/0/4373a639", "title": "Depth Representation Method by Color Tone for 3D Graphics Modeler", "doi": null, "abstractUrl": "/proceedings-article/cisis/2011/4373a639/12OmNwNOaRD", "parentPublication": { "id": "proceedings/cisis/2011/4373/0", "title": "2011 International Conference on Complex, Intelligent, and Software Intensive Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icacc/2009/3516/0/3516a087", "title": "Input-Output Exact Feedback Linearization for Depth Positioning of Spherical URV", "doi": null, "abstractUrl": "/proceedings-article/icacc/2009/3516a087/12OmNxFJXun", "parentPublication": { "id": "proceedings/icacc/2009/3516/0", "title": "2009 International Conference on Advanced Computer Control", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cmv/2003/2001/0/20010034", "title": "Integrating Multiple 3D Views through Frame-of-Reference Interaction", "doi": null, "abstractUrl": "/proceedings-article/cmv/2003/20010034/12OmNyv7m2Q", "parentPublication": { "id": "proceedings/cmv/2003/2001/0", "title": "Coordinated and Multiple Views in Exploratory Visualization, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icinis/2012/4855/0/4855a025", "title": "Depth Camera in Human-Computer Interaction: An Overview", "doi": null, "abstractUrl": "/proceedings-article/icinis/2012/4855a025/12OmNzYwbT7", "parentPublication": { "id": "proceedings/icinis/2012/4855/0", "title": "Intelligent Networks and Intelligent Systems, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mmit/2008/3556/0/3556a646", "title": "Automatic Depth Map Estimation of Monocular Indoor Environments", "doi": null, "abstractUrl": "/proceedings-article/mmit/2008/3556a646/12OmNzlUKkv", "parentPublication": { "id": "proceedings/mmit/2008/3556/0", "title": "MultiMedia and Information Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2002/09/i1226", "title": "Depth Estimation from Image Structure", "doi": null, "abstractUrl": "/journal/tp/2002/09/i1226/13rRUy0qnHq", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwCJOXn", "title": "3D User Interfaces (3DUI'06)", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2006", "__typename": "ProceedingType" }, "article": { "id": "12OmNs5rkQt", "doi": "10.1109/VR.2006.66", "title": "Group Selection Techniques for Efficient 3D Modeling", "normalizedTitle": "Group Selection Techniques for Efficient 3D Modeling", "abstract": "Object selection and manipulation (e.g. moving, rotating) are the most basic tasks in 3D scene construction. While most research on selection and manipulation techniques targets single objects, we examine the concept of group selection in this paper. Group selection is often given lesser importance than single object selection, yet is vital in providing users with a way to modify larger scenes with objects which are repetitive, sequential, or otherwise inherently understood as ?belonging together? by a user. We observed users manipulating objects in 3D scenes, and while doing so, they clearly expected that objects would be grouped based on their gravitational relationship. That is, all objects that are supported by some selected object will follow the motion of the selected object when manipulated. In this paper, we present a system that efficiently supports the manipulation of groups of objects via a gravitational hierarchy. As this hierarchy is derived with a collision detector, the new grouping techniques do not require semantic or user specified information to work. The results of the evaluation show that using the gravitational hierarchy improves scene rearrangement significantly compared to conventional non-hierarchical methods. Finally, we discuss lessons learned from this study and make some suggestions on how the results can be incorporated into other systems.", "abstracts": [ { "abstractType": "Regular", "content": "Object selection and manipulation (e.g. moving, rotating) are the most basic tasks in 3D scene construction. While most research on selection and manipulation techniques targets single objects, we examine the concept of group selection in this paper. Group selection is often given lesser importance than single object selection, yet is vital in providing users with a way to modify larger scenes with objects which are repetitive, sequential, or otherwise inherently understood as ?belonging together? by a user. We observed users manipulating objects in 3D scenes, and while doing so, they clearly expected that objects would be grouped based on their gravitational relationship. That is, all objects that are supported by some selected object will follow the motion of the selected object when manipulated. In this paper, we present a system that efficiently supports the manipulation of groups of objects via a gravitational hierarchy. As this hierarchy is derived with a collision detector, the new grouping techniques do not require semantic or user specified information to work. The results of the evaluation show that using the gravitational hierarchy improves scene rearrangement significantly compared to conventional non-hierarchical methods. Finally, we discuss lessons learned from this study and make some suggestions on how the results can be incorporated into other systems.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Object selection and manipulation (e.g. moving, rotating) are the most basic tasks in 3D scene construction. While most research on selection and manipulation techniques targets single objects, we examine the concept of group selection in this paper. Group selection is often given lesser importance than single object selection, yet is vital in providing users with a way to modify larger scenes with objects which are repetitive, sequential, or otherwise inherently understood as ?belonging together? by a user. We observed users manipulating objects in 3D scenes, and while doing so, they clearly expected that objects would be grouped based on their gravitational relationship. That is, all objects that are supported by some selected object will follow the motion of the selected object when manipulated. In this paper, we present a system that efficiently supports the manipulation of groups of objects via a gravitational hierarchy. As this hierarchy is derived with a collision detector, the new grouping techniques do not require semantic or user specified information to work. The results of the evaluation show that using the gravitational hierarchy improves scene rearrangement significantly compared to conventional non-hierarchical methods. Finally, we discuss lessons learned from this study and make some suggestions on how the results can be incorporated into other systems.", "fno": "02250095", "keywords": [ "Group Selection And Manipulation" ], "authors": [ { "affiliation": "University of Arizona, Tucson", "fullName": "Ji-Young Oh", "givenName": "Ji-Young", "surname": "Oh", "__typename": "ArticleAuthorType" }, { "affiliation": "York University Toronto, ON, Canada", "fullName": "Wolfgang Stuerzlinger", "givenName": "Wolfgang", "surname": "Stuerzlinger", "__typename": "ArticleAuthorType" }, { "affiliation": "York University Toronto, ON, Canada", "fullName": "Darius Dadgari", "givenName": "Darius", "surname": "Dadgari", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2006-03-01T00:00:00", "pubType": "proceedings", "pages": "95-102", "year": "2006", "issn": null, "isbn": "1-4244-0225-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "02250087", "articleId": "12OmNzayN6A", "__typename": "AdjacentArticleType" }, "next": { "fno": "02250103", "articleId": "12OmNwF0BWC", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iita/2008/3497/1/3497a530", "title": "Machine Tools Selection Technology for Networked Manufacturing", "doi": null, "abstractUrl": "/proceedings-article/iita/2008/3497a530/12OmNB8TU8u", "parentPublication": { "id": "proceedings/iita/2008/3497/3", "title": "2008 Second International Symposium on Intelligent Information Technology Application", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/uic-atc/2009/3737/0/3737a568", "title": "Enhancing Cooperative Behavior for Online Reputation Systems by Group Selection", "doi": null, "abstractUrl": "/proceedings-article/uic-atc/2009/3737a568/12OmNBWzHQC", "parentPublication": { "id": "proceedings/uic-atc/2009/3737/0", "title": "Ubiquitous, Autonomic and Trusted Computing, Symposia and Workshops on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444783", "title": "GPU implementation of 3D object selection by conic volume techniques in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444783/12OmNBp52w1", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/waina/2011/4338/0/4338a253", "title": "One Size Does Not Fit All: A Group-Based Service Selection for Web-Based Business Processes", "doi": null, "abstractUrl": "/proceedings-article/waina/2011/4338a253/12OmNrAMEOv", "parentPublication": { "id": "proceedings/waina/2011/4338/0", "title": "2011 IEEE Workshops of International Conference on Advanced Information Networking and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sccc/2010/4400/0/4400a218", "title": "Selection Techniques for Dense and Occluded Virtual 3D Environments, Supported by Depth Feedback: Double, Bound and Depth Bubble Cursors", "doi": null, "abstractUrl": "/proceedings-article/sccc/2010/4400a218/12OmNrAdsv3", "parentPublication": { "id": "proceedings/sccc/2010/4400/0", "title": "2010 XXIX International Conference of the Chilean Computer Science Society", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cisis/2008/3109/0/3109a951", "title": "Optimizing Decentralized Grid Markets through Group Selection", "doi": null, "abstractUrl": "/proceedings-article/cisis/2008/3109a951/12OmNrkT7Lu", "parentPublication": { "id": "proceedings/cisis/2008/3109/0", "title": "2008 International Conference on Complex, Intelligent and Software Intensive Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-infovis/1995/7201/0/72010036", "title": "SDM: malleable information graphics", "doi": null, "abstractUrl": "/proceedings-article/ieee-infovis/1995/72010036/12OmNwDSdgd", "parentPublication": { "id": "proceedings/ieee-infovis/1995/7201/0", "title": "Information Visualization, IEEE Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscid/2009/3865/2/3865b529", "title": "Research on Supplier Selection Based on Fuzzy Sets Group Decision", "doi": null, "abstractUrl": "/proceedings-article/iscid/2009/3865b529/12OmNynsbxE", "parentPublication": { "id": "proceedings/iscid/2009/3865/2", "title": "Computational Intelligence and Design, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/03/ttg2012030369", "title": "Integrality and Separability of Multitouch Interaction Techniques in 3D Manipulation Tasks", "doi": null, "abstractUrl": "/journal/tg/2012/03/ttg2012030369/13rRUx0gepW", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040634", "title": "Dense and Dynamic 3D Selection for Game-Based Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040634/13rRUyYSWsS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwCJOXn", "title": "3D User Interfaces (3DUI'06)", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2006", "__typename": "ProceedingType" }, "article": { "id": "12OmNwF0BWC", "doi": "10.1109/VR.2006.134", "title": "Towards a General Model for Selection in Virtual Environments", "normalizedTitle": "Towards a General Model for Selection in Virtual Environments", "abstract": "Selection is one of the fundamental building blocks of all interactive virtual environment systems. Selection is the ability of the user to specify which object, or sub-part of an object in the environment, is the target for subsequent actions. Examples include selecting 3D buttons thus invoking an action or selecting a target upon which an action will occur. Selection is also an implicit or explicit part of manipulation techniques. In a virtual environment selection can be performed in many different ways. In this paper we develop a generalized model of how interaction is and could be performed in virtual environments using 3D gestures. The purpose of this model is to highlight some potential areas for development and evaluation of novel selection techniques. The model is based on an analysis of the complexity of selection. We develop a model for selection that is based on time-varying scalar fields (TVSFs) that encompasses a very broad range of existing techniques. This model will be abstract, in that a direct implementation will be prohibitively complex, but we show how some standard implementation strategies are good approximations to the formal model.", "abstracts": [ { "abstractType": "Regular", "content": "Selection is one of the fundamental building blocks of all interactive virtual environment systems. Selection is the ability of the user to specify which object, or sub-part of an object in the environment, is the target for subsequent actions. Examples include selecting 3D buttons thus invoking an action or selecting a target upon which an action will occur. Selection is also an implicit or explicit part of manipulation techniques. In a virtual environment selection can be performed in many different ways. In this paper we develop a generalized model of how interaction is and could be performed in virtual environments using 3D gestures. The purpose of this model is to highlight some potential areas for development and evaluation of novel selection techniques. The model is based on an analysis of the complexity of selection. We develop a model for selection that is based on time-varying scalar fields (TVSFs) that encompasses a very broad range of existing techniques. This model will be abstract, in that a direct implementation will be prohibitively complex, but we show how some standard implementation strategies are good approximations to the formal model.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Selection is one of the fundamental building blocks of all interactive virtual environment systems. Selection is the ability of the user to specify which object, or sub-part of an object in the environment, is the target for subsequent actions. Examples include selecting 3D buttons thus invoking an action or selecting a target upon which an action will occur. Selection is also an implicit or explicit part of manipulation techniques. In a virtual environment selection can be performed in many different ways. In this paper we develop a generalized model of how interaction is and could be performed in virtual environments using 3D gestures. The purpose of this model is to highlight some potential areas for development and evaluation of novel selection techniques. The model is based on an analysis of the complexity of selection. We develop a model for selection that is based on time-varying scalar fields (TVSFs) that encompasses a very broad range of existing techniques. This model will be abstract, in that a direct implementation will be prohibitively complex, but we show how some standard implementation strategies are good approximations to the formal model.", "fno": "02250103", "keywords": [ "Selection", "3 D Interaction", "Virtual Environments" ], "authors": [ { "affiliation": "University College London", "fullName": "Anthony Steed", "givenName": "Anthony", "surname": "Steed", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2006-03-01T00:00:00", "pubType": "proceedings", "pages": "103-110", "year": "2006", "issn": null, "isbn": "1-4244-0225-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "02250095", "articleId": "12OmNs5rkQt", "__typename": "AdjacentArticleType" }, "next": { "fno": "02250111", "articleId": "12OmNx6g6dA", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dui/2006/0225/0/02250053", "title": "Using the Non-Dominant Hand for Selection in 3D", "doi": null, "abstractUrl": "/proceedings-article/3dui/2006/02250053/12OmNBZpH8h", "parentPublication": { "id": "proceedings/3dui/2006/0225/0", "title": "3D User Interfaces (3DUI'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2010/6237/0/05444783", "title": "GPU implementation of 3D object selection by conic volume techniques in virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444783/12OmNBp52w1", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/2004/2171/0/21710540", "title": "The Complexity of Testing a Motivational Model of Action Selection for Virtual Humans", "doi": null, "abstractUrl": "/proceedings-article/cgi/2004/21710540/12OmNC4eSHe", "parentPublication": { "id": "proceedings/cgi/2004/2171/0", "title": "Proceedings. Computer Graphics International", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sccc/2010/4400/0/4400a218", "title": "Selection Techniques for Dense and Occluded Virtual 3D Environments, Supported by Depth Feedback: Double, Bound and Depth Bubble Cursors", "doi": null, "abstractUrl": "/proceedings-article/sccc/2010/4400a218/12OmNrAdsv3", "parentPublication": { "id": "proceedings/sccc/2010/4400/0", "title": "2010 XXIX International Conference of the Chilean Computer Science Society", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/2001/1007/0/10070031", "title": "Towards Real Time Virtual Human Life Simulations", "doi": null, "abstractUrl": "/proceedings-article/cgi/2001/10070031/12OmNvTBB1o", "parentPublication": { "id": "proceedings/cgi/2001/1007/0", "title": "Proceedings. Computer Graphics International 2001", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2000/0478/0/04780281", "title": "Multimodal Menu Presentation and Selection in Immersive Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2000/04780281/12OmNzDehfH", "parentPublication": { "id": "proceedings/vr/2000/0478/0", "title": "Virtual Reality Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2007/0907/0/04142843", "title": "Optical Sight Metaphor for Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142843/12OmNzdoMlq", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040634", "title": "Dense and Dynamic 3D Selection for Game-Based Virtual Environments", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040634/13rRUyYSWsS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2022/8402/0/840200a916", "title": "SPE Selection Technique: A Projection-based Approach for Precise Object Interaction in Immersive Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/vrw/2022/840200a916/1CJfekhdkmQ", "parentPublication": { "id": "proceedings/vrw/2022/8402/0", "title": "2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089442", "title": "Effects of Virtual Hand Representation on Interaction and Embodiment in HMD-based Virtual Environments Using Controllers", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089442/1jIxe7ldiE0", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqMPfSj", "title": "2007 IEEE Symposium on 3D User Interfaces", "acronym": "3dui", "groupId": "1001623", "volume": "0", "displayVolume": "0", "year": "2007", "__typename": "ProceedingType" }, "article": { "id": "12OmNwK7o7o", "doi": "10.1109/3DUI.2007.340783", "title": "Exploring the Effects of Environment Density and Target Visibility on Object Selection in 3D Virtual Environments", "normalizedTitle": "Exploring the Effects of Environment Density and Target Visibility on Object Selection in 3D Virtual Environments", "abstract": "Object selection is a primary interaction technique which must be supported by any interactive three-dimensional virtual reality application. Although numerous techniques exist, few have been designed to support the selection of objects in dense target environments, or the selection of objects which are occluded from the user's viewpoint. There is, thus, a limited understanding on how these important factors will affect selection performance. In this paper, we present a set of design guidelines and strategies to aid the development of selection techniques which can compensate for environment density and target visibility. Based on these guidelines, we present two techniques, the depth ray and the 3D bubble cursor, both augmented to allow for the selection of fully occluded targets. In a formal experiment, we evaluate the relative performance of these techniques, varying both the environment density and target visibility. The results found that both of these techniques outperformed a baseline point cursor technique, with the depth ray performing best overall.", "abstracts": [ { "abstractType": "Regular", "content": "Object selection is a primary interaction technique which must be supported by any interactive three-dimensional virtual reality application. Although numerous techniques exist, few have been designed to support the selection of objects in dense target environments, or the selection of objects which are occluded from the user's viewpoint. There is, thus, a limited understanding on how these important factors will affect selection performance. In this paper, we present a set of design guidelines and strategies to aid the development of selection techniques which can compensate for environment density and target visibility. Based on these guidelines, we present two techniques, the depth ray and the 3D bubble cursor, both augmented to allow for the selection of fully occluded targets. In a formal experiment, we evaluate the relative performance of these techniques, varying both the environment density and target visibility. The results found that both of these techniques outperformed a baseline point cursor technique, with the depth ray performing best overall.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Object selection is a primary interaction technique which must be supported by any interactive three-dimensional virtual reality application. Although numerous techniques exist, few have been designed to support the selection of objects in dense target environments, or the selection of objects which are occluded from the user's viewpoint. There is, thus, a limited understanding on how these important factors will affect selection performance. In this paper, we present a set of design guidelines and strategies to aid the development of selection techniques which can compensate for environment density and target visibility. Based on these guidelines, we present two techniques, the depth ray and the 3D bubble cursor, both augmented to allow for the selection of fully occluded targets. In a formal experiment, we evaluate the relative performance of these techniques, varying both the environment density and target visibility. The results found that both of these techniques outperformed a baseline point cursor technique, with the depth ray performing best overall.", "fno": "04142854", "keywords": [ "3 D Bubble Cursor", "Environment Density", "Target Visibility", "Object Selection", "Interactive 3 D Virtual Reality", "Depth Ray" ], "authors": [ { "affiliation": "Expertise Centre for Digital Media, Hasselt Univ., Diepenbeek", "fullName": "L. Vanacken", "givenName": "L.", "surname": "Vanacken", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "T. Grossman", "givenName": "T.", "surname": "Grossman", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "K. Coninx", "givenName": "K.", "surname": "Coninx", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dui", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2007-03-01T00:00:00", "pubType": "proceedings", "pages": "null", "year": "2007", "issn": null, "isbn": "1-4244-0907-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "04142853", "articleId": "12OmNBVIUzj", "__typename": "AdjacentArticleType" }, "next": { "fno": "04142855", "articleId": "12OmNBWi6KF", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/sccc/2010/4400/0/4400a218", "title": "Selection Techniques for Dense and Occluded Virtual 3D Environments, Supported by Depth Feedback: Double, Bound and Depth Bubble Cursors", "doi": null, "abstractUrl": "/proceedings-article/sccc/2010/4400a218/12OmNrAdsv3", "parentPublication": { "id": "proceedings/sccc/2010/4400/0", "title": "2010 XXIX International Conference of the Chilean Computer Science Society", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2010/4109/0/4109b040", "title": "View-Invariant Object Recognition with Visibility Maps", "doi": null, "abstractUrl": "/proceedings-article/icpr/2010/4109b040/12OmNvIfDRE", "parentPublication": { "id": "proceedings/icpr/2010/4109/0", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csse/2008/3336/2/3336c874", "title": "An Improved LOS Method for Implementing Visibility Analysis of 3D Complex Landscapes", "doi": null, "abstractUrl": "/proceedings-article/csse/2008/3336c874/12OmNwEJ0JN", "parentPublication": { "id": "proceedings/csse/2008/3336/6", "title": "Computer Science and Software Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ds-rt/2012/4846/0/4846a019", "title": "3D Road Environment Modeling Applied to Visibility Mapping: An Experimental Comparison", "doi": null, "abstractUrl": "/proceedings-article/ds-rt/2012/4846a019/12OmNwekjIQ", "parentPublication": { "id": "proceedings/ds-rt/2012/4846/0", "title": "Distributed Simulation and Real Time Applications, IEEE/ACM International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vv/1998/9180/0/91800087", "title": "An Exact Interactive Time Visibility Ordering Algorithm for Polyhedral Cell Complexes", "doi": null, "abstractUrl": "/proceedings-article/vv/1998/91800087/12OmNxA3YRD", "parentPublication": { "id": "proceedings/vv/1998/9180/0", "title": "Volume Visualization and Graphics, IEEE Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/kse/2011/4567/0/4567a190", "title": "Some Advanced Techniques in Reducing Time for Path Planning Based on Visibility Graph", "doi": null, "abstractUrl": "/proceedings-article/kse/2011/4567a190/12OmNxFsmJ3", "parentPublication": { "id": "proceedings/kse/2011/4567/0", "title": "Knowledge and Systems Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2004/01/v0058", "title": "Binary-Space-Partitioned Images for Resolving Image-Based Visibility", "doi": null, "abstractUrl": "/journal/tg/2004/01/v0058/13rRUwdrdSp", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2001/04/v0365", "title": "Efficient Conservative Visibility Culling Using the Prioritized-Layered Projection Algorithm", "doi": null, "abstractUrl": "/journal/tg/2001/04/v0365/13rRUwgQpDd", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2003/03/v0412", "title": "A Survey of Visibility for Walkthrough Applications", "doi": null, "abstractUrl": "/journal/tg/2003/03/v0412/13rRUwh80H1", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2011/04/ttg2011040475", "title": "Analyzing Visibility Configurations", "doi": null, "abstractUrl": "/journal/tg/2011/04/ttg2011040475/13rRUy3gn7t", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNrNh0uC", "title": "2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission", "acronym": "3dimpvt", "groupId": "1800494", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNBOCWs8", "doi": "10.1109/3DIMPVT.2012.57", "title": "Accurate Full Body Scanning from a Single Fixed 3D Camera", "normalizedTitle": "Accurate Full Body Scanning from a Single Fixed 3D Camera", "abstract": "3D body modeling has been a long studied topic in computer vision and computer graphics. While several solutions have been proposed using either multiple sensors or a moving sensor, we propose here an approach when the user turns, in a natural motion, in front of a fixed 3D low cost camera. This opens the door to a wide range of applications where scanning is performed at home. Our scanning system can be easily set up and the instructions are straightforward to follow. We propose an articulated, part-based cylindrical representation for the body model, and show that accurate 3D shape can be automatically estimated from 4 key views detected from a depth video sequence. The registration between 4 key views is performed in a top-bottom-top manner which fully considers the kinematic constraints. We validate our approach on a large number of users, and compare accuracy to that of a reference laser scan. We show that even using a simplified model (5 cylinders) an average error of 5mm can be consistently achieved.", "abstracts": [ { "abstractType": "Regular", "content": "3D body modeling has been a long studied topic in computer vision and computer graphics. While several solutions have been proposed using either multiple sensors or a moving sensor, we propose here an approach when the user turns, in a natural motion, in front of a fixed 3D low cost camera. This opens the door to a wide range of applications where scanning is performed at home. Our scanning system can be easily set up and the instructions are straightforward to follow. We propose an articulated, part-based cylindrical representation for the body model, and show that accurate 3D shape can be automatically estimated from 4 key views detected from a depth video sequence. The registration between 4 key views is performed in a top-bottom-top manner which fully considers the kinematic constraints. We validate our approach on a large number of users, and compare accuracy to that of a reference laser scan. We show that even using a simplified model (5 cylinders) an average error of 5mm can be consistently achieved.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "3D body modeling has been a long studied topic in computer vision and computer graphics. While several solutions have been proposed using either multiple sensors or a moving sensor, we propose here an approach when the user turns, in a natural motion, in front of a fixed 3D low cost camera. This opens the door to a wide range of applications where scanning is performed at home. Our scanning system can be easily set up and the instructions are straightforward to follow. We propose an articulated, part-based cylindrical representation for the body model, and show that accurate 3D shape can be automatically estimated from 4 key views detected from a depth video sequence. The registration between 4 key views is performed in a top-bottom-top manner which fully considers the kinematic constraints. We validate our approach on a large number of users, and compare accuracy to that of a reference laser scan. We show that even using a simplified model (5 cylinders) an average error of 5mm can be consistently achieved.", "fno": "4873a432", "keywords": [ "Kinect", "Body Scanning", "Cylindrical Representation", "Key Pose Detection" ], "authors": [ { "affiliation": null, "fullName": "Ruizhe Wang", "givenName": "Ruizhe", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jongmoo Choi", "givenName": "Jongmoo", "surname": "Choi", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Gerard Medioni", "givenName": "Gerard", "surname": "Medioni", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dimpvt", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-10-01T00:00:00", "pubType": "proceedings", "pages": "432-439", "year": "2012", "issn": null, "isbn": "978-1-4673-4470-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4873a424", "articleId": "12OmNxA3YTw", "__typename": "AdjacentArticleType" }, "next": { "fno": "4873a440", "articleId": "12OmNqC2uVn", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dim/1997/7943/0/79430262", "title": "Reducing Movement Artifacts in Whole Body Scanning", "doi": null, "abstractUrl": "/proceedings-article/3dim/1997/79430262/12OmNAle6As", "parentPublication": { "id": "proceedings/3dim/1997/7943/0", "title": "3D Digital Imaging and Modeling, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dim/1997/7943/0/79430266", "title": "Optimal Postures and Positioning for Human Body Scanning", "doi": null, "abstractUrl": "/proceedings-article/3dim/1997/79430266/12OmNBdJ5hx", "parentPublication": { "id": "proceedings/3dim/1997/7943/0", "title": "3D Digital Imaging and Modeling, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iitaw/2009/3860/0/3860a356", "title": "An Automatic System to Reconstruct and Repair Telmat Scanning Model", "doi": null, "abstractUrl": "/proceedings-article/iitaw/2009/3860a356/12OmNrIaeh1", "parentPublication": { "id": "proceedings/iitaw/2009/3860/0", "title": "2009 Third International Symposium on Intelligent Information Technology Application Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dimpvt/2012/4873/0/4873a192", "title": "Markerless Motion Capture of Human Body Using PSO with Single Depth Camera", "doi": null, "abstractUrl": "/proceedings-article/3dimpvt/2012/4873a192/12OmNvqEvJl", "parentPublication": { "id": "proceedings/3dimpvt/2012/4873/0", "title": "2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fskd/2009/3735/7/05359994", "title": "Fuzzy Optimisation Modelling for Apparel Fit from Body Scanning Data Mining", "doi": null, "abstractUrl": "/proceedings-article/fskd/2009/05359994/12OmNyRg4dT", "parentPublication": { "id": "proceedings/fskd/2009/3735/7", "title": "Fuzzy Systems and Knowledge Discovery, Fourth International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/avss/2011/0844/0/06027284", "title": "Combined estimation of location and body pose in surveillance video", "doi": null, "abstractUrl": "/proceedings-article/avss/2011/06027284/12OmNzDvSo8", "parentPublication": { "id": "proceedings/avss/2011/0844/0", "title": "2011 8th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040643", "title": "Scanning 3D Full Human Bodies Using Kinects", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040643/13rRUwjGoFW", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2013/05/ttp2013051039", "title": "Algorithms for 3D Shape Scanning with a Depth Camera", "doi": null, "abstractUrl": "/journal/tp/2013/05/ttp2013051039/13rRUxNW1UZ", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/co/2007/07/r7028", "title": "3D Body Scanning and Healthcare Applications", "doi": null, "abstractUrl": "/magazine/co/2007/07/r7028/13rRUypGGeq", "parentPublication": { "id": "mags/co", "title": "Computer", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccmso/2022/3288/0/328800a352", "title": "Advanced 3D body scanning techniques and its clinical applications", "doi": null, "abstractUrl": "/proceedings-article/iccmso/2022/328800a352/1Mq12vt23S0", "parentPublication": { "id": "proceedings/iccmso/2022/3288/0", "title": "2022 International Conference on Computational Modelling, Simulation and Optimization (ICCMSO)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqMPfRr", "title": "3D Digital Imaging and Modeling, International Conference on", "acronym": "3dim", "groupId": "1800494", "volume": "0", "displayVolume": "0", "year": "1997", "__typename": "ProceedingType" }, "article": { "id": "12OmNBdJ5hx", "doi": "10.1109/IM.1997.603875", "title": "Optimal Postures and Positioning for Human Body Scanning", "normalizedTitle": "Optimal Postures and Positioning for Human Body Scanning", "abstract": "Abstract: Advancements in technology for digitizing the surface of the human body are providing new opportunities for research in engineering anthropometry, the study of human body measurement for design and evaluation purposes. The availability of the technology is just the first step in applying surface scanning to engineering anthropometry; several issues remain to be resolved to make these tools useful for engineering applications. One important issue is the standardization of positioning and the posture of the subject for scanning. In engineering it is not enough to be able to measure one individual one time in one posture, but it is also necessary to measure the individual in different postures and compare the individual with many other people who have been comparably measured. Not surprisingly, people can be more difficult to measure precisely than fixed stationary objects. In the process of developing standardized procedures for surveying the civilian populations of North America and Europe, an experiment was conducted to determine optimal scanning positions. While this experiment used just one type of scanning technology, many of the methods are transferable to other methods as well. This paper discusses the results from that investigation.", "abstracts": [ { "abstractType": "Regular", "content": "Abstract: Advancements in technology for digitizing the surface of the human body are providing new opportunities for research in engineering anthropometry, the study of human body measurement for design and evaluation purposes. The availability of the technology is just the first step in applying surface scanning to engineering anthropometry; several issues remain to be resolved to make these tools useful for engineering applications. One important issue is the standardization of positioning and the posture of the subject for scanning. In engineering it is not enough to be able to measure one individual one time in one posture, but it is also necessary to measure the individual in different postures and compare the individual with many other people who have been comparably measured. Not surprisingly, people can be more difficult to measure precisely than fixed stationary objects. In the process of developing standardized procedures for surveying the civilian populations of North America and Europe, an experiment was conducted to determine optimal scanning positions. While this experiment used just one type of scanning technology, many of the methods are transferable to other methods as well. This paper discusses the results from that investigation.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Abstract: Advancements in technology for digitizing the surface of the human body are providing new opportunities for research in engineering anthropometry, the study of human body measurement for design and evaluation purposes. The availability of the technology is just the first step in applying surface scanning to engineering anthropometry; several issues remain to be resolved to make these tools useful for engineering applications. One important issue is the standardization of positioning and the posture of the subject for scanning. In engineering it is not enough to be able to measure one individual one time in one posture, but it is also necessary to measure the individual in different postures and compare the individual with many other people who have been comparably measured. Not surprisingly, people can be more difficult to measure precisely than fixed stationary objects. In the process of developing standardized procedures for surveying the civilian populations of North America and Europe, an experiment was conducted to determine optimal scanning positions. While this experiment used just one type of scanning technology, many of the methods are transferable to other methods as well. This paper discusses the results from that investigation.", "fno": "79430266", "keywords": [ "Anthropology Optimal Postures Positioning Human Body Scanning Engineering Anthropometry Human Body Measurement" ], "authors": [ { "affiliation": "Sytronics Inc., Dayton, OH, USA", "fullName": "M.A. Brunsman", "givenName": "M.A.", "surname": "Brunsman", "__typename": "ArticleAuthorType" }, { "affiliation": "Sytronics Inc., Dayton, OH, USA", "fullName": "H.M. Daanen", "givenName": "H.M.", "surname": "Daanen", "__typename": "ArticleAuthorType" }, { "affiliation": "Sytronics Inc., Dayton, OH, USA", "fullName": "K.M. Robinette", "givenName": "K.M.", "surname": "Robinette", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dim", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "1997-05-01T00:00:00", "pubType": "proceedings", "pages": "266", "year": "1997", "issn": null, "isbn": "0-8186-7943-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "79430262", "articleId": "12OmNAle6As", "__typename": "AdjacentArticleType" }, "next": { "fno": "79430275", "articleId": "12OmNC1GudX", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dim/1997/7943/0/79430262", "title": "Reducing Movement Artifacts in Whole Body Scanning", "doi": null, "abstractUrl": "/proceedings-article/3dim/1997/79430262/12OmNAle6As", "parentPublication": { "id": "proceedings/3dim/1997/7943/0", "title": "3D Digital Imaging and Modeling, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dimpvt/2012/4873/0/4873a432", "title": "Accurate Full Body Scanning from a Single Fixed 3D Camera", "doi": null, "abstractUrl": "/proceedings-article/3dimpvt/2012/4873a432/12OmNBOCWs8", "parentPublication": { "id": "proceedings/3dimpvt/2012/4873/0", "title": "2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fg/1996/7713/0/77130170", "title": "Robust classification of hand postures against complex backgrounds", "doi": null, "abstractUrl": "/proceedings-article/fg/1996/77130170/12OmNBgQFSs", "parentPublication": { "id": "proceedings/fg/1996/7713/0", "title": "Proceedings of the Second International Conference on Automatic Face and Gesture Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2008/3381/0/3381a335", "title": "Automatic Surface Scanning of 3D Artifacts", "doi": null, "abstractUrl": "/proceedings-article/cw/2008/3381a335/12OmNvRU0nM", "parentPublication": { "id": "proceedings/cw/2008/3381/0", "title": "2008 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icacte/2008/3489/0/3489a246", "title": "Development of a Robotic 3D Scanning System for Reverse Engineering of Freeform Part", "doi": null, "abstractUrl": "/proceedings-article/icacte/2008/3489a246/12OmNvwTGEo", "parentPublication": { "id": "proceedings/icacte/2008/3489/0", "title": "2008 International Conference on Advanced Computer Theory and Engineering (ICACTE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2010/4215/0/4215a391", "title": "Geodesic Model of Human Body", "doi": null, "abstractUrl": "/proceedings-article/cw/2010/4215a391/12OmNwdbV6W", "parentPublication": { "id": "proceedings/cw/2010/4215/0", "title": "2010 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ccgrid/2012/4691/0/4691a912", "title": "Human Postures Recognition Based on D-S Evidence Theory and Multi-sensor Data Fusion", "doi": null, "abstractUrl": "/proceedings-article/ccgrid/2012/4691a912/12OmNy3RRFW", "parentPublication": { "id": "proceedings/ccgrid/2012/4691/0", "title": "Cluster Computing and the Grid, IEEE International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cso/2009/3605/1/3605a603", "title": "Spiral Scanning Path Scheme for Laser Rapid Prototyping", "doi": null, "abstractUrl": "/proceedings-article/cso/2009/3605a603/12OmNzTYBTF", "parentPublication": { "id": "cso/2009/3605/1", "title": "2009 International Joint Conference on Computational Sciences and Optimization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040643", "title": "Scanning 3D Full Human Bodies Using Kinects", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040643/13rRUwjGoFW", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/co/2007/07/r7028", "title": "3D Body Scanning and Healthcare Applications", "doi": null, "abstractUrl": "/magazine/co/2007/07/r7028/13rRUypGGeq", "parentPublication": { "id": "mags/co", "title": "Computer", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyPQ4vO", "title": "2008 International Conference on Cyberworlds", "acronym": "cw", "groupId": "1000175", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNvRU0nM", "doi": "10.1109/CW.2008.15", "title": "Automatic Surface Scanning of 3D Artifacts", "normalizedTitle": "Automatic Surface Scanning of 3D Artifacts", "abstract": "This paper describes an automatic 3D surface scanning technique using a 3D scanner. It allows the acquisition of a complete surface model of a 3D artifact without any manual registration and human interference. A two-pass approach is applied using a rotary table. In the first pass, a sequence of 2D images of the artifact are collected using a small rotation step. An image similarity measure is then taken to compare adjacent images to assess the differences between consecutive images to establish the optimal scanning angles. In the second pass, 3D scans are taken using the scanning angles derived in the first pass. An Iterative Closest Point (ICP) algorithm is employed to provide precise data alignment of neighboring scans. A new space encoding algorithm is developed to filter the over sampling areas for smoother polygonization. This technique is being implemented in a digital library project for the reconstructions of 3D computer models for artifacts from a collection in several Museums.", "abstracts": [ { "abstractType": "Regular", "content": "This paper describes an automatic 3D surface scanning technique using a 3D scanner. It allows the acquisition of a complete surface model of a 3D artifact without any manual registration and human interference. A two-pass approach is applied using a rotary table. In the first pass, a sequence of 2D images of the artifact are collected using a small rotation step. An image similarity measure is then taken to compare adjacent images to assess the differences between consecutive images to establish the optimal scanning angles. In the second pass, 3D scans are taken using the scanning angles derived in the first pass. An Iterative Closest Point (ICP) algorithm is employed to provide precise data alignment of neighboring scans. A new space encoding algorithm is developed to filter the over sampling areas for smoother polygonization. This technique is being implemented in a digital library project for the reconstructions of 3D computer models for artifacts from a collection in several Museums.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper describes an automatic 3D surface scanning technique using a 3D scanner. It allows the acquisition of a complete surface model of a 3D artifact without any manual registration and human interference. A two-pass approach is applied using a rotary table. In the first pass, a sequence of 2D images of the artifact are collected using a small rotation step. An image similarity measure is then taken to compare adjacent images to assess the differences between consecutive images to establish the optimal scanning angles. In the second pass, 3D scans are taken using the scanning angles derived in the first pass. An Iterative Closest Point (ICP) algorithm is employed to provide precise data alignment of neighboring scans. A new space encoding algorithm is developed to filter the over sampling areas for smoother polygonization. This technique is being implemented in a digital library project for the reconstructions of 3D computer models for artifacts from a collection in several Museums.", "fno": "3381a335", "keywords": [ "Scanning", "Image Similarity", "Digital Library", "3 D Artifacts", "Image Alignment" ], "authors": [ { "affiliation": null, "fullName": "Shiaofen Fang", "givenName": "Shiaofen", "surname": "Fang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Basil George", "givenName": "Basil", "surname": "George", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Mathew Palakal", "givenName": "Mathew", "surname": "Palakal", "__typename": "ArticleAuthorType" } ], "idPrefix": "cw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-09-01T00:00:00", "pubType": "proceedings", "pages": "335-341", "year": "2008", "issn": null, "isbn": "978-0-7695-3381-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3381a320", "articleId": "12OmNxj23aX", "__typename": "AdjacentArticleType" }, "next": { "fno": "3381a342", "articleId": "12OmNwDj1c6", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2008/2242/0/04587836", "title": "Modulated phase-shifting for 3D scanning", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2008/04587836/12OmNANBZtn", "parentPublication": { "id": "proceedings/cvpr/2008/2242/0", "title": "2008 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dim/1997/7943/0/79430262", "title": "Reducing Movement Artifacts in Whole Body Scanning", "doi": null, "abstractUrl": "/proceedings-article/3dim/1997/79430262/12OmNAle6As", "parentPublication": { "id": "proceedings/3dim/1997/7943/0", "title": "3D Digital Imaging and Modeling, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dim/2005/2327/0/23270310", "title": "Accuracy of 3D Scanning Technologies in a Face Scanning Scenario", "doi": null, "abstractUrl": "/proceedings-article/3dim/2005/23270310/12OmNCcKQxL", "parentPublication": { "id": "proceedings/3dim/2005/2327/0", "title": "3D Digital Imaging and Modeling, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dim/2007/2939/0/29390099", "title": "Interactive Scanning of Haptic Textures and Surface Compliance", "doi": null, "abstractUrl": "/proceedings-article/3dim/2007/29390099/12OmNCgrDc9", "parentPublication": { "id": "proceedings/3dim/2007/2939/0", "title": "2007 6th International Conference on 3-D Digital Imaging and Modeling", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icbmi/2011/4623/0/4623a156", "title": "3D Ultrasound Data Acqusition System Based on Back End Scan Mode", "doi": null, "abstractUrl": "/proceedings-article/icbmi/2011/4623a156/12OmNy3RRDW", "parentPublication": { "id": "proceedings/icbmi/2011/4623/0", "title": "Intelligent Computation and Bio-Medical Instrumentation, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2010/6984/0/05540082", "title": "3D shape scanning with a time-of-flight camera", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2010/05540082/12OmNzBwGH0", "parentPublication": { "id": "proceedings/cvpr/2010/6984/0", "title": "2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/smi/2007/2815/0/28150137", "title": "Automatic multiview quadruple alignment of unordered range scans", "doi": null, "abstractUrl": "/proceedings-article/smi/2007/28150137/12OmNzVoBJo", "parentPublication": { "id": "proceedings/smi/2007/2815/0", "title": "Shape Modeling and Applications, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/04/ttg2012040643", "title": "Scanning 3D Full Human Bodies Using Kinects", "doi": null, "abstractUrl": "/journal/tg/2012/04/ttg2012040643/13rRUwjGoFW", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2013/05/ttp2013051039", "title": "Algorithms for 3D Shape Scanning with a Depth Camera", "doi": null, "abstractUrl": "/journal/tp/2013/05/ttp2013051039/13rRUxNW1UZ", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/co/2007/07/r7028", "title": "3D Body Scanning and Healthcare Applications", "doi": null, "abstractUrl": "/magazine/co/2007/07/r7028/13rRUypGGeq", "parentPublication": { "id": "mags/co", "title": "Computer", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNrNh0uC", "title": "2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission", "acronym": "3dimpvt", "groupId": "1800494", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNvqEvJl", "doi": "10.1109/3DIMPVT.2012.21", "title": "Markerless Motion Capture of Human Body Using PSO with Single Depth Camera", "normalizedTitle": "Markerless Motion Capture of Human Body Using PSO with Single Depth Camera", "abstract": "This paper presents a novel approach to model human body, recover and track its 3D position, orientation and articulated-skeleton pose parameters from a single depth camera video sequence observed by Kinect sensor. In our work, human body is modeled as assembled 3D geometric primitives whose dimensions are estimated automatically. Motion parameters are recovered by projecting hypothesized body model pose to camera imaging space and seeking for optimal solution that best matches camera observation as well as physical constraints. An objective function is designed to quantify the discrepancy between the predicted and the actual, observed features and penalize implausible or unnatural pose. We exploit body skeleton's tree structure and propose a self-adaptive version of Particle Swarm Optimization (PSO) to solve the optimization problem. In order to avoid swarm collapse and accelerate convergence, motion temporal continuity over frame sequence is exploited as initial pose using from-coarse-to-fine strategy. The overall system does not require any markers, special capture environment or complex image acquisition setup, and is ready-to-use for users.", "abstracts": [ { "abstractType": "Regular", "content": "This paper presents a novel approach to model human body, recover and track its 3D position, orientation and articulated-skeleton pose parameters from a single depth camera video sequence observed by Kinect sensor. In our work, human body is modeled as assembled 3D geometric primitives whose dimensions are estimated automatically. Motion parameters are recovered by projecting hypothesized body model pose to camera imaging space and seeking for optimal solution that best matches camera observation as well as physical constraints. An objective function is designed to quantify the discrepancy between the predicted and the actual, observed features and penalize implausible or unnatural pose. We exploit body skeleton's tree structure and propose a self-adaptive version of Particle Swarm Optimization (PSO) to solve the optimization problem. In order to avoid swarm collapse and accelerate convergence, motion temporal continuity over frame sequence is exploited as initial pose using from-coarse-to-fine strategy. The overall system does not require any markers, special capture environment or complex image acquisition setup, and is ready-to-use for users.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper presents a novel approach to model human body, recover and track its 3D position, orientation and articulated-skeleton pose parameters from a single depth camera video sequence observed by Kinect sensor. In our work, human body is modeled as assembled 3D geometric primitives whose dimensions are estimated automatically. Motion parameters are recovered by projecting hypothesized body model pose to camera imaging space and seeking for optimal solution that best matches camera observation as well as physical constraints. An objective function is designed to quantify the discrepancy between the predicted and the actual, observed features and penalize implausible or unnatural pose. We exploit body skeleton's tree structure and propose a self-adaptive version of Particle Swarm Optimization (PSO) to solve the optimization problem. In order to avoid swarm collapse and accelerate convergence, motion temporal continuity over frame sequence is exploited as initial pose using from-coarse-to-fine strategy. The overall system does not require any markers, special capture environment or complex image acquisition setup, and is ready-to-use for users.", "fno": "4873a192", "keywords": [ "Self Adaptive PSO", "Human Body Motion Capture", "Marker Less", "Single Kinect" ], "authors": [ { "affiliation": null, "fullName": "Tianwei Xing", "givenName": "Tianwei", "surname": "Xing", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yao Yu", "givenName": "Yao", "surname": "Yu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yu Zhou", "givenName": "Yu", "surname": "Zhou", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Sidan Du", "givenName": "Sidan", "surname": "Du", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dimpvt", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-10-01T00:00:00", "pubType": "proceedings", "pages": "192-197", "year": "2012", "issn": null, "isbn": "978-1-4673-4470-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4873a184", "articleId": "12OmNx6xHm5", "__typename": "AdjacentArticleType" }, "next": { "fno": "4873a198", "articleId": "12OmNrAMEM1", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/wiamis/2009/3609/0/05031420", "title": "Markerless human motion capture and pose recognition", "doi": null, "abstractUrl": "/proceedings-article/wiamis/2009/05031420/12OmNANkoko", "parentPublication": { "id": "proceedings/wiamis/2009/3609/0", "title": "Image Analysis for Multimedia Interactive Services, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2007/1016/0/04284846", "title": "Model-Based Markerless Human Body Motion Capture using Multiple Cameras", "doi": null, "abstractUrl": "/proceedings-article/icme/2007/04284846/12OmNvmXJ37", "parentPublication": { "id": "proceedings/icme/2007/1016/0", "title": "2007 International Conference on Multimedia & Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/crv/2016/2491/0/2491a024", "title": "Real-Time Human Motion Capture with Multiple Depth Cameras", "doi": null, "abstractUrl": "/proceedings-article/crv/2016/2491a024/12OmNwBT1rW", "parentPublication": { "id": "proceedings/crv/2016/2491/0", "title": "2016 13th Conference on Computer and Robot Vision (CRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cmsp/2011/4356/1/4356a281", "title": "Marker-Less Human Body Tracking Using Locally Affine Invariant Contour Matching", "doi": null, "abstractUrl": "/proceedings-article/cmsp/2011/4356a281/12OmNx7XGZy", "parentPublication": { "id": "proceedings/cmsp/2011/4356/1", "title": "Multimedia and Signal Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpads/2012/4903/0/4903a284", "title": "CUDA Acceleration of 3D Dynamic Scene Reconstruction and 3D Motion Estimation for Motion Capture", "doi": null, "abstractUrl": "/proceedings-article/icpads/2012/4903a284/12OmNxETahN", "parentPublication": { "id": "proceedings/icpads/2012/4903/0", "title": "Parallel and Distributed Systems, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2013/2840/0/2840b105", "title": "Real-Time Body Tracking with One Depth Camera and Inertial Sensors", "doi": null, "abstractUrl": "/proceedings-article/iccv/2013/2840b105/12OmNyyO8GX", "parentPublication": { "id": "proceedings/iccv/2013/2840/0", "title": "2013 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2018/8425/0/842500a474", "title": "Rethinking Pose in 3D: Multi-stage Refinement and Recovery for Markerless Motion Capture", "doi": null, "abstractUrl": "/proceedings-article/3dv/2018/842500a474/17D45VN31gb", "parentPublication": { "id": "proceedings/3dv/2018/8425/0", "title": "2018 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/crv/2018/6481/0/648101a158", "title": "Walking on Thin Air: Environment-Free Physics-Based Markerless Motion Capture", "doi": null, "abstractUrl": "/proceedings-article/crv/2018/648101a158/17D45XoXP3X", "parentPublication": { "id": "proceedings/crv/2018/6481/0", "title": "2018 15th Conference on Computer and Robot Vision (CRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300k0957", "title": "Monocular Total Capture: Posing Face, Body, and Hands in the Wild", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300k0957/1gys3pjGQfe", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300a823", "title": "Markerless Outdoor Human Motion Capture Using Multiple Autonomous Micro Aerial Vehicles", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300a823/1hQqk33280w", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwt5sgJ", "title": "CVPR 2011", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNwBjP7F", "doi": "10.1109/CVPR.2011.5995321", "title": "Structured light 3D scanning in the presence of global illumination", "normalizedTitle": "Structured light 3D scanning in the presence of global illumination", "abstract": "Global illumination effects such as inter-reflections, diffusion and sub-surface scattering severely degrade the performance of structured light-based 3D scanning. In this paper, we analyze the errors caused by global illumination in structured light-based shape recovery. Based on this analysis, we design structured light patterns that are resilient to individual global illumination effects using simple logical operations and tools from combinatorial mathematics. Scenes exhibiting multiple phenomena are handled by combining results from a small ensemble of such patterns. This combination also allows us to detect any residual errors that are corrected by acquiring a few additional images. Our techniques do not require explicit separation of the direct and global components of scene radiance and hence work even in scenarios where the separation fails or the direct component is too low. Our methods can be readily incorporated into existing scanning systems without significant overhead in terms of capture time or hardware. We show results on a variety of scenes with complex shape and material properties and challenging global illumination effects.", "abstracts": [ { "abstractType": "Regular", "content": "Global illumination effects such as inter-reflections, diffusion and sub-surface scattering severely degrade the performance of structured light-based 3D scanning. In this paper, we analyze the errors caused by global illumination in structured light-based shape recovery. Based on this analysis, we design structured light patterns that are resilient to individual global illumination effects using simple logical operations and tools from combinatorial mathematics. Scenes exhibiting multiple phenomena are handled by combining results from a small ensemble of such patterns. This combination also allows us to detect any residual errors that are corrected by acquiring a few additional images. Our techniques do not require explicit separation of the direct and global components of scene radiance and hence work even in scenarios where the separation fails or the direct component is too low. Our methods can be readily incorporated into existing scanning systems without significant overhead in terms of capture time or hardware. We show results on a variety of scenes with complex shape and material properties and challenging global illumination effects.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Global illumination effects such as inter-reflections, diffusion and sub-surface scattering severely degrade the performance of structured light-based 3D scanning. In this paper, we analyze the errors caused by global illumination in structured light-based shape recovery. Based on this analysis, we design structured light patterns that are resilient to individual global illumination effects using simple logical operations and tools from combinatorial mathematics. Scenes exhibiting multiple phenomena are handled by combining results from a small ensemble of such patterns. This combination also allows us to detect any residual errors that are corrected by acquiring a few additional images. Our techniques do not require explicit separation of the direct and global components of scene radiance and hence work even in scenarios where the separation fails or the direct component is too low. Our methods can be readily incorporated into existing scanning systems without significant overhead in terms of capture time or hardware. We show results on a variety of scenes with complex shape and material properties and challenging global illumination effects.", "fno": "05995321", "keywords": [ "Scanning Systems", "Structured Light", "3 D Scanning", "Global Illumination", "Shape Recovery" ], "authors": [ { "affiliation": "Robot. Inst., Carnegie Mellon Univ., Pittsburgh, PA, USA", "fullName": "M. Gupta", "givenName": "M.", "surname": "Gupta", "__typename": "ArticleAuthorType" }, { "affiliation": "Mitsubishi Electr. Res. Labs., Cambridge, MA, USA", "fullName": "A. Agrawal", "givenName": "A.", "surname": "Agrawal", "__typename": "ArticleAuthorType" }, { "affiliation": "Mitsubishi Electr. Res. Labs., Cambridge, MA, USA", "fullName": "A. Veeraraghavan", "givenName": "A.", "surname": "Veeraraghavan", "__typename": "ArticleAuthorType" }, { "affiliation": "Robot. Inst., Carnegie Mellon Univ., Pittsburgh, PA, USA", "fullName": "S. G. Narasimhan", "givenName": "S. G.", "surname": "Narasimhan", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-06-01T00:00:00", "pubType": "proceedings", "pages": "713-720", "year": "2011", "issn": null, "isbn": "978-1-4577-0394-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05995320", "articleId": "12OmNyUFfYY", "__typename": "AdjacentArticleType" }, "next": { "fno": "05995322", "articleId": "12OmNwqft52", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/aici/2010/4225/2/4225b063", "title": "Multi-resolution Decoding Method of Symbol M Array Surface Structured Light", "doi": null, "abstractUrl": "/proceedings-article/aici/2010/4225b063/12OmNBBhN6R", "parentPublication": { "id": "proceedings/aici/2010/4225/2", "title": "Artificial Intelligence and Computational Intelligence, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dim/1997/7943/0/79430028", "title": "Toward Optimal Structured Light Patterns", "doi": null, "abstractUrl": "/proceedings-article/3dim/1997/79430028/12OmNBKmXpP", "parentPublication": { "id": "proceedings/3dim/1997/7943/0", "title": "3D Digital Imaging and Modeling, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/psivt/2010/4285/0/4285a276", "title": "Modeling and Detection of Blurred Illumination Edges", "doi": null, "abstractUrl": "/proceedings-article/psivt/2010/4285a276/12OmNBf94VX", "parentPublication": { "id": "proceedings/psivt/2010/4285/0", "title": "Image and Video Technology, Pacific-Rim Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2011/1101/0/06126305", "title": "Multiplexed illumination for scene recovery in the presence of global illumination", "doi": null, "abstractUrl": "/proceedings-article/iccv/2011/06126305/12OmNrAv3TW", "parentPublication": { "id": "proceedings/iccv/2011/1101/0", "title": "2011 IEEE International Conference on Computer Vision (ICCV 2011)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dimpvt/2011/4369/0/4369a389", "title": "Temporally-Consistent Phase Unwrapping for a Stereo-Assisted Structured Light System", "doi": null, "abstractUrl": "/proceedings-article/3dimpvt/2011/4369a389/12OmNwDACD1", "parentPublication": { "id": "proceedings/3dimpvt/2011/4369/0", "title": "2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dimpvt/2011/4369/0/4369a397", "title": "A Multi-camera, Multi-projector Super-Resolution Framework for Structured Light", "doi": null, "abstractUrl": "/proceedings-article/3dimpvt/2011/4369a397/12OmNxcMSiL", "parentPublication": { "id": "proceedings/3dimpvt/2011/4369/0", "title": "2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dimpvt/2012/4873/0/4873a456", "title": "Sparse Modeling of Shape from Structured Light", "doi": null, "abstractUrl": "/proceedings-article/3dimpvt/2012/4873a456/12OmNy2rS5x", "parentPublication": { "id": "proceedings/3dimpvt/2012/4873/0", "title": "2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2010/4109/0/4109a798", "title": "A Color Invariant Based Binary Coded Structured Light Range Scanner for Shiny Objects", "doi": null, "abstractUrl": "/proceedings-article/icpr/2010/4109a798/12OmNySXEW3", "parentPublication": { "id": "proceedings/icpr/2010/4109/0", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2010/4109/0/4109b650", "title": "Extending Fast Marching Method under Point Light Source Illumination and Perspective Projection", "doi": null, "abstractUrl": "/proceedings-article/icpr/2010/4109b650/12OmNyz5JTC", "parentPublication": { "id": "proceedings/icpr/2010/4109/0", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2009/06/ttg2009061595", "title": "Decoupling Illumination from Isosurface Generation Using 4D Light Transport", "doi": null, "abstractUrl": "/journal/tg/2009/06/ttg2009061595/13rRUwInvf0", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAolGIV", "title": "Intelligent Computation and Bio-Medical Instrumentation, International Conference on", "acronym": "icbmi", "groupId": "1800786", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNy3RRDW", "doi": "10.1109/ICBMI.2011.1", "title": "3D Ultrasound Data Acqusition System Based on Back End Scan Mode", "normalizedTitle": "3D Ultrasound Data Acqusition System Based on Back End Scan Mode", "abstract": "The data acquisition of 3D-Ultrasound includes array scan and mechanical scan, and the later one is more easy to realize. Currently, the traditional probe scanning mode is Front-end scanning. Under the above scanning mode, when it scans over the breast, 2D-Ultrasound probe slides through the surface of the patients' bodies, the image will be influenced strongly by the human ribs.In this paper we propose a new scanning mode to solve the above problem: Back-End scan, the back end rotates while the front end contacts the patients' skin without slide. The device designed using dual stepper motors which are under the synchronization control. Experiment results show the effectiveness of the proposed device.", "abstracts": [ { "abstractType": "Regular", "content": "The data acquisition of 3D-Ultrasound includes array scan and mechanical scan, and the later one is more easy to realize. Currently, the traditional probe scanning mode is Front-end scanning. Under the above scanning mode, when it scans over the breast, 2D-Ultrasound probe slides through the surface of the patients' bodies, the image will be influenced strongly by the human ribs.In this paper we propose a new scanning mode to solve the above problem: Back-End scan, the back end rotates while the front end contacts the patients' skin without slide. The device designed using dual stepper motors which are under the synchronization control. Experiment results show the effectiveness of the proposed device.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The data acquisition of 3D-Ultrasound includes array scan and mechanical scan, and the later one is more easy to realize. Currently, the traditional probe scanning mode is Front-end scanning. Under the above scanning mode, when it scans over the breast, 2D-Ultrasound probe slides through the surface of the patients' bodies, the image will be influenced strongly by the human ribs.In this paper we propose a new scanning mode to solve the above problem: Back-End scan, the back end rotates while the front end contacts the patients' skin without slide. The device designed using dual stepper motors which are under the synchronization control. Experiment results show the effectiveness of the proposed device.", "fno": "4623a156", "keywords": [ "Sector Scanning Mode", "Back End Scanning", "3 D Reconstruction" ], "authors": [ { "affiliation": null, "fullName": "Jian Qi", "givenName": "Jian", "surname": "Qi", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Mingyue Ding", "givenName": "Mingyue", "surname": "Ding", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ming Yuchi", "givenName": "Ming", "surname": "Yuchi", "__typename": "ArticleAuthorType" } ], "idPrefix": "icbmi", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-12-01T00:00:00", "pubType": "proceedings", "pages": "156-158", "year": "2011", "issn": null, "isbn": "978-0-7695-4623-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4623a152", "articleId": "12OmNwtn3D6", "__typename": "AdjacentArticleType" }, "next": { "fno": "4623a159", "articleId": "12OmNro0Ie5", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icmtma/2011/4296/1/4296a766", "title": "Construction of 3D Finite Element Model of Upper End of Tibia", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2011/4296a766/12OmNB1eJyL", "parentPublication": { "id": "proceedings/icmtma/2011/4296/1", "title": "2011 Third International Conference on Measuring Technology and Mechatronics Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csse/2008/3336/2/3336d012", "title": "Research on a New Kind of Adaptive Parallel Scan Method in Laser Metal Deposition Shaping", "doi": null, "abstractUrl": "/proceedings-article/csse/2008/3336d012/12OmNBEGYLh", "parentPublication": { "id": "proceedings/csse/2008/3336/6", "title": "Computer Science and Software Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cdciem/2012/4639/0/4639a393", "title": "LP0: A RFID Authentication Protocol for Low-cost Tags without Back-end Database", "doi": null, "abstractUrl": "/proceedings-article/cdciem/2012/4639a393/12OmNButq5D", "parentPublication": { "id": "proceedings/cdciem/2012/4639/0", "title": "Computer Distributed Control and Intelligent Environmental Monitoring, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/1994/6275/0/00577143", "title": "Radial homomorphic deconvolution of B-mode medical ultrasound images", "doi": null, "abstractUrl": "/proceedings-article/icpr/1994/00577143/12OmNweTvNY", "parentPublication": { "id": "proceedings/icpr/1994/6275/0", "title": "12th IAPR International Conference on Pattern Recognition, 1994", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mmbia/2001/1336/0/13360063", "title": "Deformation Correction in Ultrasound Images Using Contact Force Measurements", "doi": null, "abstractUrl": "/proceedings-article/mmbia/2001/13360063/12OmNyxFKkq", "parentPublication": { "id": "proceedings/mmbia/2001/1336/0", "title": "Workshop on Mathematical Methods in Biomedical Image Analysis", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icia/2006/0528/0/04097986", "title": "Kriging Interpolation Based Ultrasound Scan Conversion Algorithm", "doi": null, "abstractUrl": "/proceedings-article/icia/2006/04097986/12OmNz6iOCb", "parentPublication": { "id": "proceedings/icia/2006/0528/0", "title": "2006 International Conference on Information Acquisition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/itc/2014/4722/0/07035357", "title": "Test-mode-only scan attack and countermeasure for contemporary scan architectures", "doi": null, "abstractUrl": "/proceedings-article/itc/2014/07035357/12OmNzuIjlL", "parentPublication": { "id": "proceedings/itc/2014/4722/0", "title": "2014 IEEE International Test Conference (ITC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/dt/2007/06/mdt2007060546", "title": "Secured CAD Back-End Flow for Power-Analysis-Resistant Cryptoprocessors", "doi": null, "abstractUrl": "/magazine/dt/2007/06/mdt2007060546/13rRUxNmPJY", "parentPublication": { "id": "mags/dt", "title": "IEEE Design & Test of Computers", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vlsid/2022/8505/0/850500a011", "title": "A PC based Ultrasound back-end signal processor using Intel&#x00AE; Performance Primitives", "doi": null, "abstractUrl": "/proceedings-article/vlsid/2022/850500a011/1GFaCUBAioU", "parentPublication": { "id": "proceedings/vlsid/2022/8505/0", "title": "2022 35th International Conference on VLSI Design and 2022 21st International Conference on Embedded Systems (VLSID)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzuIjee", "title": "Digital Media and Digital Content Management, Workshop on", "acronym": "dmdcm", "groupId": "1800440", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNznCkZN", "doi": "10.1109/DMDCM.2011.26", "title": "A Study on Natural Interaction for Human Body Motion Using Depth Image Data", "normalizedTitle": "A Study on Natural Interaction for Human Body Motion Using Depth Image Data", "abstract": "This paper explains a study on natural interaction (NI) in human body motion using depth image data. It involves about overview of NI and depth image data. Human body motion is a non-verbal part for interaction or movement that can be used to involves real world and virtual world. Furthermore, interaction with computer or machine can be more realistic as real world and becoming more important to academic researchers, game industries, and can be adapt to other field like mechanical engineering for robotics movement and surgery purpose in medical area. Functional taxonomies will show step-by-step how human body motion were detected and created a skeleton joint. Also, we discuss about technologies behind Kinect for Xbox 360 (Kinect). Recent research in this area also included.", "abstracts": [ { "abstractType": "Regular", "content": "This paper explains a study on natural interaction (NI) in human body motion using depth image data. It involves about overview of NI and depth image data. Human body motion is a non-verbal part for interaction or movement that can be used to involves real world and virtual world. Furthermore, interaction with computer or machine can be more realistic as real world and becoming more important to academic researchers, game industries, and can be adapt to other field like mechanical engineering for robotics movement and surgery purpose in medical area. Functional taxonomies will show step-by-step how human body motion were detected and created a skeleton joint. Also, we discuss about technologies behind Kinect for Xbox 360 (Kinect). Recent research in this area also included.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper explains a study on natural interaction (NI) in human body motion using depth image data. It involves about overview of NI and depth image data. Human body motion is a non-verbal part for interaction or movement that can be used to involves real world and virtual world. Furthermore, interaction with computer or machine can be more realistic as real world and becoming more important to academic researchers, game industries, and can be adapt to other field like mechanical engineering for robotics movement and surgery purpose in medical area. Functional taxonomies will show step-by-step how human body motion were detected and created a skeleton joint. Also, we discuss about technologies behind Kinect for Xbox 360 (Kinect). Recent research in this area also included.", "fno": "4413a097", "keywords": [ "Human Computer Interaction", "Natural Interaction", "Natural User Interface", "Human Body Motion", "Depth Image Data", "Depth Aware Camera", "Kinect" ], "authors": [ { "affiliation": null, "fullName": "Mohd Kufaisal bin Mohd Sidik", "givenName": "Mohd Kufaisal bin Mohd", "surname": "Sidik", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Mohd Shahrizal bin Sunar", "givenName": "Mohd Shahrizal bin", "surname": "Sunar", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ismahafezi bin Ismail", "givenName": "Ismahafezi bin", "surname": "Ismail", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Mohd Khalid bin Mokhtar", "givenName": "Mohd Khalid bin", "surname": "Mokhtar", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Normal binti Mat Jusoh", "givenName": "Normal binti Mat", "surname": "Jusoh", "__typename": "ArticleAuthorType" } ], "idPrefix": "dmdcm", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-05-01T00:00:00", "pubType": "proceedings", "pages": "97-102", "year": "2011", "issn": null, "isbn": "978-0-7695-4413-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4413a091", "articleId": "12OmNCcKQz4", "__typename": "AdjacentArticleType" }, "next": { "fno": "4413a103", "articleId": "12OmNvSKNXf", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/mvhi/2010/4009/0/4009a247", "title": "Motion Control Based on Dimensional Reduction and Human Computer Interaction", "doi": null, "abstractUrl": "/proceedings-article/mvhi/2010/4009a247/12OmNBCqbCT", "parentPublication": { "id": "proceedings/mvhi/2010/4009/0", "title": "Machine Vision and Human-machine Interface, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dimpvt/2012/4873/0/4873a192", "title": "Markerless Motion Capture of Human Body Using PSO with Single Depth Camera", "doi": null, "abstractUrl": "/proceedings-article/3dimpvt/2012/4873a192/12OmNvqEvJl", "parentPublication": { "id": "proceedings/3dimpvt/2012/4873/0", "title": "2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2012/4725/0/4725a191", "title": "Guidance and Movement Correction Based on Therapeutics Movements for Motor Rehabilitation Support Systems", "doi": null, "abstractUrl": "/proceedings-article/svr/2012/4725a191/12OmNwtWfRu", "parentPublication": { "id": "proceedings/svr/2012/4725/0", "title": "2012 14th Symposium on Virtual and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/crv/2012/4683/0/4683a460", "title": "In Situ Motion Capture of Speed Skating: Escaping the Treadmill", "doi": null, "abstractUrl": "/proceedings-article/crv/2012/4683a460/12OmNxwncya", "parentPublication": { "id": "proceedings/crv/2012/4683/0", "title": "2012 Ninth Conference on Computer and Robot Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2011/4445/0/4445a112", "title": "Interaction in Augmented Reality Environments Using Kinect", "doi": null, "abstractUrl": "/proceedings-article/svr/2011/4445a112/12OmNy6HQOK", "parentPublication": { "id": "proceedings/svr/2011/4445/0", "title": "2011 XIII Symposium on Virtual Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/svr/2012/4725/0/4725a271", "title": "Human Body Motion and Gestures Recognition Based on Checkpoints", "doi": null, "abstractUrl": "/proceedings-article/svr/2012/4725a271/12OmNyQph2l", "parentPublication": { "id": "proceedings/svr/2012/4725/0", "title": "2012 14th Symposium on Virtual and Augmented Reality", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cimsim/2011/4562/0/4562a248", "title": "Tracking of Fingertips and Centers of Palm Using KINECT", "doi": null, "abstractUrl": "/proceedings-article/cimsim/2011/4562a248/12OmNz4BdsP", "parentPublication": { "id": "proceedings/cimsim/2011/4562/0", "title": "Computational Intelligence, Modelling and Simulation, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/12/ttg2012122719", "title": "Interaction Support for Visual Comparison Inspired by Natural Behavior", "doi": null, "abstractUrl": "/journal/tg/2012/12/ttg2012122719/13rRUxZRbo0", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/mu/2012/02/mmu2012020004", "title": "Microsoft Kinect Sensor and Its Effect", "doi": null, "abstractUrl": "/magazine/mu/2012/02/mmu2012020004/13rRUxly9aE", "parentPublication": { "id": "mags/mu", "title": "IEEE MultiMedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2016/08/07312488", "title": "Probabilistic Social Behavior Analysis by Exploring Body Motion-Based Patterns", "doi": null, "abstractUrl": "/journal/tp/2016/08/07312488/13rRUyYBlhT", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxA3Z41", "title": "2010 Third International Conference on Information and Computing", "acronym": "icic", "groupId": "1002818", "volume": "2", "displayVolume": "2", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNzw8jgm", "doi": "10.1109/ICIC.2010.145", "title": "Fitting the Sectional Curve for 3D Human Body Based on Non-uniform B-spline", "normalizedTitle": "Fitting the Sectional Curve for 3D Human Body Based on Non-uniform B-spline", "abstract": "The main object of the research is about 3D human body. Firstly, obtaining data of standard mannequin by [TC]2 3D body scanner, then using the steps of non-uniform B-spine such as derivation, inverse calculating, control points to analyse the data, at last editing programs by matlab to receive the fitting chart of 3D human body. The result is establishing elements of the size of 3D human body.", "abstracts": [ { "abstractType": "Regular", "content": "The main object of the research is about 3D human body. Firstly, obtaining data of standard mannequin by [TC]2 3D body scanner, then using the steps of non-uniform B-spine such as derivation, inverse calculating, control points to analyse the data, at last editing programs by matlab to receive the fitting chart of 3D human body. The result is establishing elements of the size of 3D human body.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The main object of the research is about 3D human body. Firstly, obtaining data of standard mannequin by [TC]2 3D body scanner, then using the steps of non-uniform B-spine such as derivation, inverse calculating, control points to analyse the data, at last editing programs by matlab to receive the fitting chart of 3D human body. The result is establishing elements of the size of 3D human body.", "fno": "4047b201", "keywords": [ "3 D Human Body Section", "Non Uniform B Spline", "Fitting Curve" ], "authors": [ { "affiliation": null, "fullName": "Zi-min Jin", "givenName": "Zi-min", "surname": "Jin", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yu-xiu Yan", "givenName": "Yu-xiu", "surname": "Yan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jue-min Mu", "givenName": "Jue-min", "surname": "Mu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Li-fang Liao", "givenName": "Li-fang", "surname": "Liao", "__typename": "ArticleAuthorType" } ], "idPrefix": "icic", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-06-01T00:00:00", "pubType": "proceedings", "pages": "201-204", "year": "2010", "issn": null, "isbn": "978-0-7695-4047-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4047b198", "articleId": "12OmNCbCrHR", "__typename": "AdjacentArticleType" }, "next": { "fno": "4047b205", "articleId": "12OmNzBOin4", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/sc/1992/2630/0/00236697", "title": "Vectorized algorithm for B-spline curve fitting on CRAY X-MP EA/16se", "doi": null, "abstractUrl": "/proceedings-article/sc/1992/00236697/12OmNASILS5", "parentPublication": { "id": "proceedings/sc/1992/2630/0", "title": "SC Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccsee/2012/4647/3/4647c508", "title": "Quadratic Non-uniform Hyperbolic B-spline Curves", "doi": null, "abstractUrl": "/proceedings-article/iccsee/2012/4647c508/12OmNCbU32m", "parentPublication": { "id": "proceedings/iccsee/2012/4647/3", "title": "Computer Science and Electronics Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csse/2008/3336/2/3336c975", "title": "Uniform B-Spline Curve and Surface with Shape Parameters", "doi": null, "abstractUrl": "/proceedings-article/csse/2008/3336c975/12OmNqBbI2O", "parentPublication": { "id": "proceedings/csse/2008/3336/6", "title": "Computer Science and Software Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/gmp/2000/0562/0/05620150", "title": "Optimization of Knots for the Multi Curve B-Spline Approximation", "doi": null, "abstractUrl": "/proceedings-article/gmp/2000/05620150/12OmNxiKs6C", "parentPublication": { "id": "proceedings/gmp/2000/0562/0", "title": "Geometric Modeling and Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/kam/2009/3888/1/3888a220", "title": "Transformer Winding Vibration Enveloping for Empirical Mode Decomposition Based on Non-uniform B-Spline Fitting", "doi": null, "abstractUrl": "/proceedings-article/kam/2009/3888a220/12OmNy2rRVm", "parentPublication": { "id": null, "title": null, "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2000/0743/0/07430063", "title": "Curve Designing Using a Rational Cubic Spline with Point and Interval Shape Control", "doi": null, "abstractUrl": "/proceedings-article/iv/2000/07430063/12OmNy8prAE", "parentPublication": { "id": "proceedings/iv/2000/0743/0", "title": "2000 IEEE Conference on Information Visualization. An International Conference on Computer Visualization and Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iptc/2010/4196/0/4196a021", "title": "Tension Spline Fitting with Bezier Curve", "doi": null, "abstractUrl": "/proceedings-article/iptc/2010/4196a021/12OmNzZEAGt", "parentPublication": { "id": "proceedings/iptc/2010/4196/0", "title": "Intelligence Information Processing and Trusted Computing, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdar/2011/4520/0/4520b115", "title": "A Compression Scheme for Handwritten Patterns Based on Curve Fitting", "doi": null, "abstractUrl": "/proceedings-article/icdar/2011/4520b115/12OmNzdoMXM", "parentPublication": { "id": "proceedings/icdar/2011/4520/0", "title": "2011 International Conference on Document Analysis and Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/09/06035703", "title": "Uniform B-Spline Curve Interpolation with Prescribed Tangent and Curvature Vectors", "doi": null, "abstractUrl": "/journal/tg/2012/09/06035703/13rRUxYIN48", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/itme/2022/1015/0/101500a180", "title": "Personalized Custom Virtual Fitting Display Method", "doi": null, "abstractUrl": "/proceedings-article/itme/2022/101500a180/1M4rv5Veb0k", "parentPublication": { "id": "proceedings/itme/2022/1015/0", "title": "2022 12th International Conference on Information Technology in Medicine and Education (ITME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNCbCrVF", "title": "1999 IEEE International Conference on Information Visualization (Cat. No. PR00210)", "acronym": "iv", "groupId": "1000370", "volume": "0", "displayVolume": "0", "year": "1999", "__typename": "ProceedingType" }, "article": { "id": "12OmNAIvd0i", "doi": "10.1109/IV.1999.781580", "title": "Towards Interactive Finite Element Analysis of Shell Structures in Virtual Reality", "normalizedTitle": "Towards Interactive Finite Element Analysis of Shell Structures in Virtual Reality", "abstract": "A first step towards a semi-immersive Virtual Reality (VR) interface for Finite Element Analysis (FEA) is presented in this paper. During recent years, user interfaces of FEA solvers have matured from character- based command-line driven implementations into easy-to-use graphical user interfaces (GUIs). This new generation of GUIs provides access to intuitive and productive tools for the management and analysis of structural problems. Many pre- and post-processors have been implemented targeting the simplification of the man-machine interface in order to increase the ease of use and provide better visual analysis of FEA solver results. Nevertheless, none of these packages provides a real 3D-enabled interface. The main objective of this project is to join state-of-the-art visualization technology, VR devices, and FEA solvers into the integrated development environment VRFEA.", "abstracts": [ { "abstractType": "Regular", "content": "A first step towards a semi-immersive Virtual Reality (VR) interface for Finite Element Analysis (FEA) is presented in this paper. During recent years, user interfaces of FEA solvers have matured from character- based command-line driven implementations into easy-to-use graphical user interfaces (GUIs). This new generation of GUIs provides access to intuitive and productive tools for the management and analysis of structural problems. Many pre- and post-processors have been implemented targeting the simplification of the man-machine interface in order to increase the ease of use and provide better visual analysis of FEA solver results. Nevertheless, none of these packages provides a real 3D-enabled interface. The main objective of this project is to join state-of-the-art visualization technology, VR devices, and FEA solvers into the integrated development environment VRFEA.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "A first step towards a semi-immersive Virtual Reality (VR) interface for Finite Element Analysis (FEA) is presented in this paper. During recent years, user interfaces of FEA solvers have matured from character- based command-line driven implementations into easy-to-use graphical user interfaces (GUIs). This new generation of GUIs provides access to intuitive and productive tools for the management and analysis of structural problems. Many pre- and post-processors have been implemented targeting the simplification of the man-machine interface in order to increase the ease of use and provide better visual analysis of FEA solver results. Nevertheless, none of these packages provides a real 3D-enabled interface. The main objective of this project is to join state-of-the-art visualization technology, VR devices, and FEA solvers into the integrated development environment VRFEA.", "fno": "02100340", "keywords": [ "Virtual Reality", "Interactive Modeling", "Finite Element Analysis", "3 D Modeling", "Simulation" ], "authors": [ { "affiliation": "University of Bologna", "fullName": "A. Liverani", "givenName": "A.", "surname": "Liverani", "__typename": "ArticleAuthorType" }, { "affiliation": "University of California at Davis", "fullName": "F. Kuester", "givenName": "F.", "surname": "Kuester", "__typename": "ArticleAuthorType" }, { "affiliation": "University of California at Davis", "fullName": "B. Hamann", "givenName": "B.", "surname": "Hamann", "__typename": "ArticleAuthorType" } ], "idPrefix": "iv", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1999-07-01T00:00:00", "pubType": "proceedings", "pages": "340", "year": "1999", "issn": "1093-9547", "isbn": "0-7695-0210-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "02100334", "articleId": "12OmNAoDhUT", "__typename": "AdjacentArticleType" }, "next": { "fno": "02100348", "articleId": "12OmNrJ11Ix", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvRU0le", "title": "Geometric Modeling and Processing", "acronym": "gmp", "groupId": "1000306", "volume": "0", "displayVolume": "0", "year": "2000", "__typename": "ProceedingType" }, "article": { "id": "12OmNBbaH9z", "doi": "10.1109/GMAP.2000.838236", "title": "Rapid 3D Model Acquisition from Images of Small Objects", "normalizedTitle": "Rapid 3D Model Acquisition from Images of Small Objects", "abstract": "This paper describes a system for rapid acquisition of 3D models of objects for use in applications such as CAD and VR. The system consists of an image capturing platform, which comprises a PC-controlled turntable and CCD camera, and associated computer vision and graphics algorithms for 3D model acquisition. An algorithm for accurately recovering 3D points from image sequences is described. Experiments performed on synthetic and real data show that the algorithm is accurate in recovering the coordinates of 3D points and is robust against noise in 2D feature location and 3D object rotation. Results of applying the system to acquiring 3D models of real objects are illustrated.", "abstracts": [ { "abstractType": "Regular", "content": "This paper describes a system for rapid acquisition of 3D models of objects for use in applications such as CAD and VR. The system consists of an image capturing platform, which comprises a PC-controlled turntable and CCD camera, and associated computer vision and graphics algorithms for 3D model acquisition. An algorithm for accurately recovering 3D points from image sequences is described. Experiments performed on synthetic and real data show that the algorithm is accurate in recovering the coordinates of 3D points and is robust against noise in 2D feature location and 3D object rotation. Results of applying the system to acquiring 3D models of real objects are illustrated.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper describes a system for rapid acquisition of 3D models of objects for use in applications such as CAD and VR. The system consists of an image capturing platform, which comprises a PC-controlled turntable and CCD camera, and associated computer vision and graphics algorithms for 3D model acquisition. An algorithm for accurately recovering 3D points from image sequences is described. Experiments performed on synthetic and real data show that the algorithm is accurate in recovering the coordinates of 3D points and is robust against noise in 2D feature location and 3D object rotation. Results of applying the system to acquiring 3D models of real objects are illustrated.", "fno": "05620033", "keywords": [ "Model Acquisition", "3 D Reconstruction", "Image Sequence", "Turntable Sequence" ], "authors": [ { "affiliation": "National University of Singapore", "fullName": "Wee Kheng Leow", "givenName": "Wee Kheng", "surname": "Leow", "__typename": "ArticleAuthorType" }, { "affiliation": "National University of Singapore", "fullName": "Zhiyong Huang", "givenName": "Zhiyong", "surname": "Huang", "__typename": "ArticleAuthorType" }, { "affiliation": "National University of Singapore", "fullName": "Yong Zhang", "givenName": "Yong", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": "National University of Singapore", "fullName": "Rudy Setiono", "givenName": "Rudy", "surname": "Setiono", "__typename": "ArticleAuthorType" } ], "idPrefix": "gmp", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "2000-04-01T00:00:00", "pubType": "proceedings", "pages": "33", "year": "2000", "issn": null, "isbn": "0-7695-0562-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05620013", "articleId": "12OmNz6iO4O", "__typename": "AdjacentArticleType" }, "next": { "fno": "05620045", "articleId": "12OmNxWcH9j", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNCy2L3V", "title": "Cyber-Physical Systems, IEEE/ACM International Conference on", "acronym": "iccps", "groupId": "1800417", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNButpWP", "doi": "10.1109/ICCPS.2012.44", "title": "Demo Abstract: Towards a Wireless Building Management System with Minimum Change to the Building Protocols", "normalizedTitle": "Demo Abstract: Towards a Wireless Building Management System with Minimum Change to the Building Protocols", "abstract": "We develop a wireless system that can (partially) substitute the wired infrastructure underlying Building Automation and Control Networks (BACnet) of the Building Management System (BMS). There are efforts (e.g. from ZigBee alliance) in developing the wireless components for BMS, they build from bottom up which may lead to change and re-standardization of a new BACnet protocol. There are also entirely new infrastructure proposed (e.g. sMAP) to collect physical information, however, these schemes require a longer time to be adopted. Our system is designed to support BACnet without any modification. We believe that our design experience is useful for other applications where wired infrastructure and upper layer protocols that are complete and well established, and is considered to be cost-effective in converting the system to (partially) wireless. Our experience shows that a direct replacement of wires into wireless links does not work. There are a number of difficulties, including the strict and harsh time constraints for frames to be transmitted before the system consider it as timeout. Our solution is to propose an asynchronous-response framework so as to keep the control flow of the upper layers protocol intact. In this demo, we will 1) bring the physical devices and explain our system on-site and 2) play a experiment video of our system that recorded in real environment.", "abstracts": [ { "abstractType": "Regular", "content": "We develop a wireless system that can (partially) substitute the wired infrastructure underlying Building Automation and Control Networks (BACnet) of the Building Management System (BMS). There are efforts (e.g. from ZigBee alliance) in developing the wireless components for BMS, they build from bottom up which may lead to change and re-standardization of a new BACnet protocol. There are also entirely new infrastructure proposed (e.g. sMAP) to collect physical information, however, these schemes require a longer time to be adopted. Our system is designed to support BACnet without any modification. We believe that our design experience is useful for other applications where wired infrastructure and upper layer protocols that are complete and well established, and is considered to be cost-effective in converting the system to (partially) wireless. Our experience shows that a direct replacement of wires into wireless links does not work. There are a number of difficulties, including the strict and harsh time constraints for frames to be transmitted before the system consider it as timeout. Our solution is to propose an asynchronous-response framework so as to keep the control flow of the upper layers protocol intact. In this demo, we will 1) bring the physical devices and explain our system on-site and 2) play a experiment video of our system that recorded in real environment.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We develop a wireless system that can (partially) substitute the wired infrastructure underlying Building Automation and Control Networks (BACnet) of the Building Management System (BMS). There are efforts (e.g. from ZigBee alliance) in developing the wireless components for BMS, they build from bottom up which may lead to change and re-standardization of a new BACnet protocol. There are also entirely new infrastructure proposed (e.g. sMAP) to collect physical information, however, these schemes require a longer time to be adopted. Our system is designed to support BACnet without any modification. We believe that our design experience is useful for other applications where wired infrastructure and upper layer protocols that are complete and well established, and is considered to be cost-effective in converting the system to (partially) wireless. Our experience shows that a direct replacement of wires into wireless links does not work. There are a number of difficulties, including the strict and harsh time constraints for frames to be transmitted before the system consider it as timeout. Our solution is to propose an asynchronous-response framework so as to keep the control flow of the upper layers protocol intact. In this demo, we will 1) bring the physical devices and explain our system on-site and 2) play a experiment video of our system that recorded in real environment.", "fno": "4695a223", "keywords": [ "BA Cnet", "BMS", "Building Management System", "Building Automation" ], "authors": [ { "affiliation": null, "fullName": "Qinghua Luo", "givenName": "Qinghua", "surname": "Luo", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Abraham Hang-Yat Lam", "givenName": "Abraham Hang-Yat", "surname": "Lam", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Dan Wang", "givenName": "Dan", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Daniel Wai-Tin Chan", "givenName": "Daniel Wai-Tin", "surname": "Chan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yu Peng", "givenName": "Yu", "surname": "Peng", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Xiyuan Peng", "givenName": "Xiyuan", "surname": "Peng", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccps", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-04-01T00:00:00", "pubType": "proceedings", "pages": "223", "year": "2012", "issn": null, "isbn": "978-0-7695-4695-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4695a222", "articleId": "12OmNvonIGL", "__typename": "AdjacentArticleType" }, "next": { "fno": "4695a224", "articleId": "12OmNwx3QcO", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icss/2013/4972/0/4972a266", "title": "Intelligent Building Automation and Control Based on IndasIBMS", "doi": null, "abstractUrl": "/proceedings-article/icss/2013/4972a266/12OmNB0nWh3", "parentPublication": { "id": "proceedings/icss/2013/4972/0", "title": "Service Sciences, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iacsit-sc/2009/3653/0/3653a626", "title": "Integrated Simulation Platform for Optimized Building Operations", "doi": null, "abstractUrl": "/proceedings-article/iacsit-sc/2009/3653a626/12OmNB1eJxr", "parentPublication": { "id": "proceedings/iacsit-sc/2009/3653/0", "title": "Computer Science and Information Technology, International Association of", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ipsn/2012/6469/0/06920928", "title": "Demo abstract: Personal building controls", "doi": null, "abstractUrl": "/proceedings-article/ipsn/2012/06920928/12OmNBpmDGx", "parentPublication": { "id": "proceedings/ipsn/2012/6469/0", "title": "2012 ACM/IEEE 11th International Conference on Information Processing in Sensor Networks (IPSN)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aina/2008/3095/0/3095a713", "title": "SARBAU ? Towards Highly Self-Configuring IP-Fieldbus Based Building Automation Networks", "doi": null, "abstractUrl": "/proceedings-article/aina/2008/3095a713/12OmNvSbBrC", "parentPublication": { "id": "proceedings/aina/2008/3095/0", "title": "22nd International Conference on Advanced Information Networking and Applications (aina 2008)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cccm/2008/3290/1/3290a528", "title": "Software Interface Technology of Intelligent Building System Integration", "doi": null, "abstractUrl": "/proceedings-article/cccm/2008/3290a528/12OmNwIHotS", "parentPublication": { "id": "cccm/2008/3290/1", "title": "Computing, Communication, Control and Management, ISECS International Colloquium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/greencom/2012/4865/0/4865a731", "title": "Covert Channels and Their Prevention in Building Automation Protocols: A Prototype Exemplified Using BACnet", "doi": null, "abstractUrl": "/proceedings-article/greencom/2012/4865a731/12OmNx8wTf8", "parentPublication": { "id": "proceedings/greencom/2012/4865/0", "title": "2012 IEEE International Conference on Green Computing and Communications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/isbim/2008/3560/2/3560b063", "title": "Design of a Web-Based Building Management System Using Ajax and Web Services", "doi": null, "abstractUrl": "/proceedings-article/isbim/2008/3560b063/12OmNxYL5fP", "parentPublication": { "id": "proceedings/isbim/2008/3560/2", "title": "Business and Information Management, International Seminar on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ipsn/2009/5108/0/05211889", "title": "Demo abstract: Embedded Virtual Machines for wireless industrial automation", "doi": null, "abstractUrl": "/proceedings-article/ipsn/2009/05211889/12OmNyGbIjr", "parentPublication": { "id": "proceedings/ipsn/2009/5108/0", "title": "2009 8th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/alpit/2007/2930/0/2930a376", "title": "Building BIS/BMS with Wireless Communication System in Korea", "doi": null, "abstractUrl": "/proceedings-article/alpit/2007/2930a376/12OmNz4SOvi", "parentPublication": { "id": "proceedings/alpit/2007/2930/0", "title": "Advanced Language Processing and Web Information Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/compeng/2010/3974/0/3974a058", "title": "A System Level Model of Possible Integration of Building Management System in SmartGrid", "doi": null, "abstractUrl": "/proceedings-article/compeng/2010/3974a058/12OmNzWfp8Z", "parentPublication": { "id": "proceedings/compeng/2010/3974/0", "title": "Engineering. Complexity in", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBqMDoJ", "title": "2008 International Symposium on Information Science and Engineering (ISISE)", "acronym": "isise", "groupId": "1002561", "volume": "2", "displayVolume": "2", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNwO5LU2", "doi": "10.1109/ISISE.2008.328", "title": "Analysis on Interactive Structure of Knowledge Acquisition", "normalizedTitle": "Analysis on Interactive Structure of Knowledge Acquisition", "abstract": "Knowledge Acquisition has been the focus in knowledge engineering for a long time. In this paper, knowledge acquisition technique based on structural analysis is formed by applying structural modeling, which is on basis of knowledge structural model. This technique can be applied to knowledge acquisition by dynamic interaction; and the new knowledge is related to the existing meta knowledge system. By enlighten of the existing knowledge, this method can automatically accomplish knowledge architecture and implicit relation acquisition. It is applied to Traditional Chinese Medicine (TCM) diagnosis and treatment knowledge acquisition in this paper; and the results show that the structural model not only indicates unknown relations, but also protrudes the structural relations of existing knowledge, thus making knowledge acquisition more efficient.", "abstracts": [ { "abstractType": "Regular", "content": "Knowledge Acquisition has been the focus in knowledge engineering for a long time. In this paper, knowledge acquisition technique based on structural analysis is formed by applying structural modeling, which is on basis of knowledge structural model. This technique can be applied to knowledge acquisition by dynamic interaction; and the new knowledge is related to the existing meta knowledge system. By enlighten of the existing knowledge, this method can automatically accomplish knowledge architecture and implicit relation acquisition. It is applied to Traditional Chinese Medicine (TCM) diagnosis and treatment knowledge acquisition in this paper; and the results show that the structural model not only indicates unknown relations, but also protrudes the structural relations of existing knowledge, thus making knowledge acquisition more efficient.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Knowledge Acquisition has been the focus in knowledge engineering for a long time. In this paper, knowledge acquisition technique based on structural analysis is formed by applying structural modeling, which is on basis of knowledge structural model. This technique can be applied to knowledge acquisition by dynamic interaction; and the new knowledge is related to the existing meta knowledge system. By enlighten of the existing knowledge, this method can automatically accomplish knowledge architecture and implicit relation acquisition. It is applied to Traditional Chinese Medicine (TCM) diagnosis and treatment knowledge acquisition in this paper; and the results show that the structural model not only indicates unknown relations, but also protrudes the structural relations of existing knowledge, thus making knowledge acquisition more efficient.", "fno": "3494b654", "keywords": [ "Interactive Structure", "Knowledge Acquisition" ], "authors": [ { "affiliation": null, "fullName": "Dezheng Zhang", "givenName": "Dezheng", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Aziguli Wulamu", "givenName": "Aziguli", "surname": "Wulamu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Lixin Gao", "givenName": "Lixin", "surname": "Gao", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Pin Shan", "givenName": "Pin", "surname": "Shan", "__typename": "ArticleAuthorType" } ], "idPrefix": "isise", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-12-01T00:00:00", "pubType": "proceedings", "pages": "654-658", "year": "2008", "issn": null, "isbn": "978-0-7695-3494-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3494b650", "articleId": "12OmNzFv4jY", "__typename": "AdjacentArticleType" }, "next": { "fno": "3494b659", "articleId": "12OmNrH1PAB", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/csse/2008/3336/5/3336i381", "title": "How Team Spirit Improve Product Innovation: The Role of Shared Vision and Knowledge Acquisition", "doi": null, "abstractUrl": "/proceedings-article/csse/2008/3336i381/12OmNqOffvE", "parentPublication": { "id": "proceedings/csse/2008/3336/5", "title": "2008 International Conference on Computer Science and Software Engineering (CSSE 2008)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iita/2008/3497/1/3497a181", "title": "Study on Knowledge Acquisition Techniques", "doi": null, "abstractUrl": "/proceedings-article/iita/2008/3497a181/12OmNvDI45r", "parentPublication": { "id": "proceedings/iita/2008/3497/3", "title": "2008 Second International Symposium on Intelligent Information Technology Application", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iciii/2008/3435/3/3435c335", "title": "An Integrated Model of Knowledge Acquisition: Empirical Evidences in China", "doi": null, "abstractUrl": "/proceedings-article/iciii/2008/3435c335/12OmNwCsdPw", "parentPublication": { "id": "proceedings/iciii/2008/3435/3", "title": "International Conference on Information Management, Innovation Management and Industrial Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/imis/2015/8873/0/8873a467", "title": "Professional Network Structure and Relationship Quality Effects on Member Knowledge Acquisition, Well-Being, and Performance", "doi": null, "abstractUrl": "/proceedings-article/imis/2015/8873a467/12OmNwE9OLw", "parentPublication": { "id": "proceedings/imis/2015/8873/0", "title": "2015 9th International Conference on Innovative Mobile and Internet Services in Ubiquitous Computing (IMIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icm/2011/4522/4/4522d093", "title": "Knowledge Transfer in Equipment Acquisition Project", "doi": null, "abstractUrl": "/proceedings-article/icm/2011/4522d093/12OmNzC5T4r", "parentPublication": { "id": "proceedings/icm/2011/4522/4", "title": "Information Technology, Computer Engineering and Management Sciences, International Conference of", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csie/2009/3507/5/3507e195", "title": "Working Structure Knowledge Acquisition from Mechanical Product Patent Based on Natural Language Understanding", "doi": null, "abstractUrl": "/proceedings-article/csie/2009/3507e195/12OmNzkuKzR", "parentPublication": { "id": "proceedings/csie/2009/3507/6", "title": "Computer Science and Information Engineering, World Congress on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icss/2015/9947/0/9947a057", "title": "The Impact of Knowledge Acquisition and Knowledge Integration of IT Outsourcing Supplier on Outsourcing Success -- Knowledge Sticky's Moderating Effect", "doi": null, "abstractUrl": "/proceedings-article/icss/2015/9947a057/12OmNzlly4i", "parentPublication": { "id": "proceedings/icss/2015/9947/0", "title": "2015 International Conference on Service Science (ICSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tk/1994/01/k0003", "title": "Document Processing for Automatic Knowledge Acquisition", "doi": null, "abstractUrl": "/journal/tk/1994/01/k0003/13rRUwh80Hn", "parentPublication": { "id": "trans/tk", "title": "IEEE Transactions on Knowledge & Data Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/so/1993/03/s3023", "title": "Metatools for Knowledge Acquisition", "doi": null, "abstractUrl": "/magazine/so/1993/03/s3023/13rRUxBJht5", "parentPublication": { "id": "mags/so", "title": "IEEE Software", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icsgea/2021/3263/0/326300a180", "title": "Automatic knowledge acquisition in interactive oral translation", "doi": null, "abstractUrl": "/proceedings-article/icsgea/2021/326300a180/1vb9jKtk7WE", "parentPublication": { "id": "proceedings/icsgea/2021/3263/0", "title": "2021 6th International Conference on Smart Grid and Electrical Automation (ICSGEA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvAiSpc", "title": "Intelligent Networks and Intelligent Systems, International Workshop on", "acronym": "icinis", "groupId": "1002524", "volume": "0", "displayVolume": "0", "year": "2008", "__typename": "ProceedingType" }, "article": { "id": "12OmNyYDDMj", "doi": "10.1109/ICINIS.2008.122", "title": "Automatic Texture Acquisition for 3D Model Using Oblique Aerial Images", "normalizedTitle": "Automatic Texture Acquisition for 3D Model Using Oblique Aerial Images", "abstract": "This paper describes an approach for automatic texture acquisition obtained from 3D city model and oblique aerial images. Firstly, the image feature lines should be extracted. Secondly, the extracted lines are matched with the corresponding 3D feature lines of object-space from the 3D city model, by estimating the image-space position of the 3D lines with the coarsely provided Exterior Orientation (EO) parameters of the image. Then, for each corresponding lines, we could list two equations based on the coplanarity constraints of object-space; that is, the accurate EO parameters can be calculated by Least Square Method (LSM) with at least three pairs of conjugate lines. Finally, the interest image areas including the building surfaces are rectified with the refined EO parameters and the textures are obtained. In this paper, the principles of the method are described in detail along with an experiment carried out on a dataset of downtown area abroad, which has proved the correctness and robustness of such method.", "abstracts": [ { "abstractType": "Regular", "content": "This paper describes an approach for automatic texture acquisition obtained from 3D city model and oblique aerial images. Firstly, the image feature lines should be extracted. Secondly, the extracted lines are matched with the corresponding 3D feature lines of object-space from the 3D city model, by estimating the image-space position of the 3D lines with the coarsely provided Exterior Orientation (EO) parameters of the image. Then, for each corresponding lines, we could list two equations based on the coplanarity constraints of object-space; that is, the accurate EO parameters can be calculated by Least Square Method (LSM) with at least three pairs of conjugate lines. Finally, the interest image areas including the building surfaces are rectified with the refined EO parameters and the textures are obtained. In this paper, the principles of the method are described in detail along with an experiment carried out on a dataset of downtown area abroad, which has proved the correctness and robustness of such method.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper describes an approach for automatic texture acquisition obtained from 3D city model and oblique aerial images. Firstly, the image feature lines should be extracted. Secondly, the extracted lines are matched with the corresponding 3D feature lines of object-space from the 3D city model, by estimating the image-space position of the 3D lines with the coarsely provided Exterior Orientation (EO) parameters of the image. Then, for each corresponding lines, we could list two equations based on the coplanarity constraints of object-space; that is, the accurate EO parameters can be calculated by Least Square Method (LSM) with at least three pairs of conjugate lines. Finally, the interest image areas including the building surfaces are rectified with the refined EO parameters and the textures are obtained. In this paper, the principles of the method are described in detail along with an experiment carried out on a dataset of downtown area abroad, which has proved the correctness and robustness of such method.", "fno": "3391a495", "keywords": [ "Texture Acquisition", "3 D Model", "Oblique Aerial Images" ], "authors": [ { "affiliation": null, "fullName": "Mi Wang", "givenName": "Mi", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Hao Bai", "givenName": "Hao", "surname": "Bai", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Fen Hu", "givenName": "Fen", "surname": "Hu", "__typename": "ArticleAuthorType" } ], "idPrefix": "icinis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2008-11-01T00:00:00", "pubType": "proceedings", "pages": "495-498", "year": "2008", "issn": null, "isbn": "978-0-7695-3391-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3391a491", "articleId": "12OmNvlPkGG", "__typename": "AdjacentArticleType" }, "next": { "fno": "3391a499", "articleId": "12OmNBAIAQZ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cisp/2008/3119/2/3119b145", "title": "Super Resolution of 3D Surface Texture Based on Eigen Images", "doi": null, "abstractUrl": "/proceedings-article/cisp/2008/3119b145/12OmNBC8AAT", "parentPublication": { "id": "proceedings/cisp/2008/3119/3", "title": "Image and Signal Processing, Congress on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dpvt/2004/2223/0/22230396", "title": "Automated Texture Mapping of 3D City Models With Oblique Aerial Imagery", "doi": null, "abstractUrl": "/proceedings-article/3dpvt/2004/22230396/12OmNBO3K4u", "parentPublication": { "id": "proceedings/3dpvt/2004/2223/0", "title": "3D Data Processing Visualization and Transmission, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscsct/2008/3498/2/3498b230", "title": "3D Surface Texture Synthesis Based on Wavelet Transform", "doi": null, "abstractUrl": "/proceedings-article/iscsct/2008/3498b230/12OmNBRsVxV", "parentPublication": { "id": "proceedings/iscsct/2008/3498/1", "title": "2008 International Symposium on Computer Science and Computational Technology (ISCSCT)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/gmp/2000/0562/0/05620033", "title": "Rapid 3D Model Acquisition from Images of Small Objects", "doi": null, "abstractUrl": "/proceedings-article/gmp/2000/05620033/12OmNBbaH9z", "parentPublication": { "id": "proceedings/gmp/2000/0562/0", "title": "Geometric Modeling and Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/1995/7042/0/70420121", "title": "Texture segmentation and shape in the same image", "doi": null, "abstractUrl": "/proceedings-article/iccv/1995/70420121/12OmNvkpl2J", "parentPublication": { "id": "proceedings/iccv/1995/7042/0", "title": "Computer Vision, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/1995/7042/0/70420888", "title": "Site model acquisition and extension from aerial images", "doi": null, "abstractUrl": "/proceedings-article/iccv/1995/70420888/12OmNxxvAIL", "parentPublication": { "id": "proceedings/iccv/1995/7042/0", "title": "Computer Vision, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/1997/7822/0/78221074", "title": "Hyper-patches for 3D model acquisition and tracking", "doi": null, "abstractUrl": "/proceedings-article/cvpr/1997/78221074/12OmNzSh18X", "parentPublication": { "id": "proceedings/cvpr/1997/7822/0", "title": "Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icip/1994/6952/1/00413307", "title": "Automatic registration of oblique aerial images", "doi": null, "abstractUrl": "/proceedings-article/icip/1994/00413307/12OmNzzfTpJ", "parentPublication": { "id": "proceedings/icip/1994/6952/3", "title": "Proceedings of 1st International Conference on Image Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2001/05/i0501", "title": "Detection and Modeling of Buildings from Multiple Aerial Images", "doi": null, "abstractUrl": "/journal/tp/2001/05/i0501/13rRUwInv5n", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/1992/11/i1111", "title": "Extraction of Straight Lines in Aerial Images", "doi": null, "abstractUrl": "/journal/tp/1992/11/i1111/13rRUxjQypU", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNscfI2d", "title": "Information Technology, Computer Engineering and Management Sciences, International Conference of", "acronym": "icm", "groupId": "1800613", "volume": "2", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNyaoDx2", "doi": "10.1109/ICM.2011.36", "title": "Component Warehouse Driven by Workflow Based on Building Block", "normalizedTitle": "Component Warehouse Driven by Workflow Based on Building Block", "abstract": "Software development based on building block (SDBBB) theory has gained increasing attention. The workflow technology is applied to software development. Workflow technology based on building block has been given, which can be applied to component-based visual programming, thus workflow technology and component warehouse technology can be combined to form a visual SDBBB mode. A workflow process model based on building block has been put forward, which is the theoretical basis for SDBBB, and is a reference model of component warehouse driven by workflow. Workflow engine system based on building block has been designed. The data model and control model for workflow engine are presented. It has been pointed out that SDBBB is not only suitable for e-government system, but also for the field of visual software development.", "abstracts": [ { "abstractType": "Regular", "content": "Software development based on building block (SDBBB) theory has gained increasing attention. The workflow technology is applied to software development. Workflow technology based on building block has been given, which can be applied to component-based visual programming, thus workflow technology and component warehouse technology can be combined to form a visual SDBBB mode. A workflow process model based on building block has been put forward, which is the theoretical basis for SDBBB, and is a reference model of component warehouse driven by workflow. Workflow engine system based on building block has been designed. The data model and control model for workflow engine are presented. It has been pointed out that SDBBB is not only suitable for e-government system, but also for the field of visual software development.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Software development based on building block (SDBBB) theory has gained increasing attention. The workflow technology is applied to software development. Workflow technology based on building block has been given, which can be applied to component-based visual programming, thus workflow technology and component warehouse technology can be combined to form a visual SDBBB mode. A workflow process model based on building block has been put forward, which is the theoretical basis for SDBBB, and is a reference model of component warehouse driven by workflow. Workflow engine system based on building block has been designed. The data model and control model for workflow engine are presented. It has been pointed out that SDBBB is not only suitable for e-government system, but also for the field of visual software development.", "fno": "4522b034", "keywords": [ "E Commerce", "Workflow", "Software Development Based On Building Block SDBBB" ], "authors": [ { "affiliation": null, "fullName": "Cheng Zhang", "givenName": "Cheng", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Huiming Tang", "givenName": "Huiming", "surname": "Tang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yiping Wu", "givenName": "Yiping", "surname": "Wu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Liangqing Wang", "givenName": "Liangqing", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Cheng Zhang", "givenName": "Cheng", "surname": "Zhang", "__typename": "ArticleAuthorType" } ], "idPrefix": "icm", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-09-01T00:00:00", "pubType": "proceedings", "pages": "34-37", "year": "2011", "issn": null, "isbn": "978-0-7695-4522-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4522b030", "articleId": "12OmNxXCGGs", "__typename": "AdjacentArticleType" }, "next": { "fno": "4522b038", "articleId": "12OmNxiKs6V", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/wkdd/2009/3543/0/3543a116", "title": "Analysis Methods of Workflow Execution Data Based on Data Mining", "doi": null, "abstractUrl": "/proceedings-article/wkdd/2009/3543a116/12OmNA0MZ9c", "parentPublication": { "id": "proceedings/wkdd/2009/3543/0", "title": "2009 Second International Workshop on Knowledge Discovery and Data Mining. WKDD 2009", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icita/2005/2316/1/231610209", "title": "Mechanism of Semantic Oriented Flexible Workflow", "doi": null, "abstractUrl": "/proceedings-article/icita/2005/231610209/12OmNAkWvFx", "parentPublication": { "id": "proceedings/icita/2005/2316/1", "title": "Proceedings. Third International Conference on Information Technology and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hicss/2001/0981/7/09817050", "title": "A Framework for Classifying Interorganizational Workflow-Controlled Business Processes Focusing on Quality Management", "doi": null, "abstractUrl": "/proceedings-article/hicss/2001/09817050/12OmNBBQZlp", "parentPublication": { "id": "proceedings/hicss/2001/0981/7", "title": "Proceedings of the 34th Annual Hawaii International Conference on System Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icee/2010/3997/0/3997e483", "title": "Research on the Land and Resources E-government System at City Level Based on Building Platform", "doi": null, "abstractUrl": "/proceedings-article/icee/2010/3997e483/12OmNqyUUH3", "parentPublication": { "id": "proceedings/icee/2010/3997/0", "title": "International Conference on E-Business and E-Government", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mines/2011/4559/0/4559a233", "title": "BSWB: To Construct Sound Workflow Model", "doi": null, "abstractUrl": "/proceedings-article/mines/2011/4559a233/12OmNwFid3l", "parentPublication": { "id": "proceedings/mines/2011/4559/0", "title": "Multimedia Information Networking and Security, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cec/2005/2277/0/22770110", "title": "Cross Enterprise Business Modeling with AC Diagrams and Workflow Patterns", "doi": null, "abstractUrl": "/proceedings-article/cec/2005/22770110/12OmNxRF6ZK", "parentPublication": { "id": "proceedings/cec/2005/2277/0", "title": "Seventh IEEE International Conference on E-Commerce Technology (CEC'05)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/isecs/2009/3643/1/3643a408", "title": "Application of GRID Computating and Workflow Technology in E-Commerce Learning", "doi": null, "abstractUrl": "/proceedings-article/isecs/2009/3643a408/12OmNy31492", "parentPublication": { "id": "proceedings/isecs/2009/3643/2", "title": "Electronic Commerce and Security, International Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icde/2002/1531/0/15310287", "title": "Integrating Workflow Management Systems with Business-to-Business Interaction Standards", "doi": null, "abstractUrl": "/proceedings-article/icde/2002/15310287/12OmNzSh1cg", "parentPublication": { "id": "proceedings/icde/2002/1531/0", "title": "Proceedings 18th International Conference on Data Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wism/2009/3817/0/3817a635", "title": "The Application of an Improved Flexibility Workflow in E-Commence", "doi": null, "abstractUrl": "/proceedings-article/wism/2009/3817a635/12OmNzTH14T", "parentPublication": { "id": "proceedings/wism/2009/3817/0", "title": "Web Information Systems and Mining, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cs/2019/04/08726147", "title": "Middleware Building Blocks for Workflow Systems", "doi": null, "abstractUrl": "/magazine/cs/2019/04/08726147/1asefnVTwZ2", "parentPublication": { "id": "mags/cs", "title": "Computing in Science & Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNy5hRcX", "title": "Information Technology and Computer Science, International Conference on", "acronym": "itcs", "groupId": "1002810", "volume": "1", "displayVolume": "1", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNz4Bdnz", "doi": "10.1109/ITCS.2009.134", "title": "Implementation of the Interactive Gestures of Virtual Avatar Based on a Multi-user Virtual Learning Environment", "normalizedTitle": "Implementation of the Interactive Gestures of Virtual Avatar Based on a Multi-user Virtual Learning Environment", "abstract": "The application of virtual worlds through Internet for teaching allows that the student can submerge himself in friendly environments that make learning more pleasant. However, the student’s avatar of the VRML plugin in Internet Explorer can not do anything without walking. This article presents an application and wants to implement the 3D virtual Avatar interaction gestures. The first, we introduce what is a virtual 3D avatar. In the following content, we build a virtual museum of history and culture as a multi-user virtual learning environment and introduce how to make a virtual 3D avatar with many interactive gestures. Finally, the latest advances in the creation of an avatar interaction server that users can share 3D interactive gestures. This has made the design of a more interactive realistic world where the student can interact with other users and share his emotions.", "abstracts": [ { "abstractType": "Regular", "content": "The application of virtual worlds through Internet for teaching allows that the student can submerge himself in friendly environments that make learning more pleasant. However, the student’s avatar of the VRML plugin in Internet Explorer can not do anything without walking. This article presents an application and wants to implement the 3D virtual Avatar interaction gestures. The first, we introduce what is a virtual 3D avatar. In the following content, we build a virtual museum of history and culture as a multi-user virtual learning environment and introduce how to make a virtual 3D avatar with many interactive gestures. Finally, the latest advances in the creation of an avatar interaction server that users can share 3D interactive gestures. This has made the design of a more interactive realistic world where the student can interact with other users and share his emotions.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The application of virtual worlds through Internet for teaching allows that the student can submerge himself in friendly environments that make learning more pleasant. However, the student’s avatar of the VRML plugin in Internet Explorer can not do anything without walking. This article presents an application and wants to implement the 3D virtual Avatar interaction gestures. The first, we introduce what is a virtual 3D avatar. In the following content, we build a virtual museum of history and culture as a multi-user virtual learning environment and introduce how to make a virtual 3D avatar with many interactive gestures. Finally, the latest advances in the creation of an avatar interaction server that users can share 3D interactive gestures. This has made the design of a more interactive realistic world where the student can interact with other users and share his emotions.", "fno": "3688a613", "keywords": [ "3 D", "Virtual Avatar", "3 D Interactive Gesture" ], "authors": [ { "affiliation": null, "fullName": "Bo Mu", "givenName": "Bo", "surname": "Mu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "YuHui Yang", "givenName": "YuHui", "surname": "Yang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "JianPing Zhang", "givenName": "JianPing", "surname": "Zhang", "__typename": "ArticleAuthorType" } ], "idPrefix": "itcs", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-07-01T00:00:00", "pubType": "proceedings", "pages": "613-617", "year": "2009", "issn": null, "isbn": "978-0-7695-3688-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3688a609", "articleId": "12OmNylboEZ", "__typename": "AdjacentArticleType" }, "next": { "fno": "3688a618", "articleId": "12OmNAndicZ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iita/2008/3497/2/3497c016", "title": "Three Dimension Human Body Format and Its Virtual Avatar Animation Application", "doi": null, "abstractUrl": "/proceedings-article/iita/2008/3497c016/12OmNBdruaX", "parentPublication": { "id": "iita/2008/3497/2", "title": "2008 Second International Symposium on Intelligent Information Technology Application", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2014/4677/0/4677a093", "title": "User Avatar Association in Virtual Worlds", "doi": null, "abstractUrl": "/proceedings-article/cw/2014/4677a093/12OmNBhHt8t", "parentPublication": { "id": "proceedings/cw/2014/4677/0", "title": "2014 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icacte/2008/3489/0/3489a286", "title": "Development and Application of Virtual Avatar Interactive Technology", "doi": null, "abstractUrl": "/proceedings-article/icacte/2008/3489a286/12OmNrHjqPb", "parentPublication": { "id": "proceedings/icacte/2008/3489/0", "title": "2008 International Conference on Advanced Computer Theory and Engineering (ICACTE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icctd/2009/3892/2/3892b416", "title": "Haptic Vibration for Emotional Expression of Avatar to Enhance the Realism of Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/icctd/2009/3892b416/12OmNs0TKTU", "parentPublication": { "id": "proceedings/icctd/2009/3892/2", "title": "Computer Technology and Development, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2010/4215/0/4215a072", "title": "Speech-Based Emotion Characterization Using Postures and Gestures in CVEs", "doi": null, "abstractUrl": "/proceedings-article/cw/2010/4215a072/12OmNvSKNN3", "parentPublication": { "id": "proceedings/cw/2010/4215/0", "title": "2010 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccsit/2008/3308/0/3308a741", "title": "Concept and Application of Three Dimension Virtual Study Technology for Academic e-Learning", "doi": null, "abstractUrl": "/proceedings-article/iccsit/2008/3308a741/12OmNvjyxKu", "parentPublication": { "id": "proceedings/iccsit/2008/3308/0", "title": "2008 International Conference on Computer Science and Information Technology", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2007/0907/0/04142838", "title": "The Visual Appearance of User?s Avatar Can Influence the Manipulation of Both Real Devices and Virtual Objects", "doi": null, "abstractUrl": "/proceedings-article/3dui/2007/04142838/12OmNwErpsL", "parentPublication": { "id": "proceedings/3dui/2007/0907/0", "title": "2007 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icece/2010/4031/0/4031a257", "title": "A Six-Degree-of-Freedom Virtual Mouse Based on Hand Gestures", "doi": null, "abstractUrl": "/proceedings-article/icece/2010/4031a257/12OmNzb7Zla", "parentPublication": { "id": "proceedings/icece/2010/4031/0", "title": "Electrical and Control Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wkdd/2009/3543/0/3543a717", "title": "Exploration of Virtual Avatar Technology Based on VRML and Java", "doi": null, "abstractUrl": "/proceedings-article/wkdd/2009/3543a717/12OmNzd7bC6", "parentPublication": { "id": "proceedings/wkdd/2009/3543/0", "title": "2009 Second International Workshop on Knowledge Discovery and Data Mining. WKDD 2009", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/ex/2010/06/mex2010060017", "title": "An Integrated Framework for Avatar Data Collection from the Virtual World", "doi": null, "abstractUrl": "/magazine/ex/2010/06/mex2010060017/13rRUB7a0WV", "parentPublication": { "id": "mags/ex", "title": "IEEE Intelligent Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAYoKlI", "title": "ASE/IEEE International Conference on BioMedical Computing (BioMedCom)", "acronym": "biomedcom", "groupId": "1802652", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNzSQdmV", "doi": "10.1109/BioMedCom.2012.28", "title": "Introducing Security Building Block Models", "normalizedTitle": "Introducing Security Building Block Models", "abstract": "In todayâs software development process, security related design decisions are rarely made early in the overall process. Even if security is considered early, this means that in most cases a more-or-less encompassing security requirement analyses is made, Based on this analysis best-practices, ad-hocdesign decisions or individual expertise is used to integrate security during the development process or after weaknesses are found after the deployment. This paper introduces Security Building Block Models which are used to build security related components, namely Security Building Blocks. These Security Building Blocks represent concrete security solutions, so called Security Properties, introduced in other publications of the Sec Futur project. The goal of this approach is to provide already defined and tested security related software components, which can be used early in the overall development process, to support security-design-decision already while modeling the software-system. The paper shortly describes this new Security Engineering Process with its requirement analysis and definition of Security Properties and how the Security Building Block Model fits into this approach. Additionally the Security Building Block Model is presented in detail. All artifacts and relationships of the model are described. Short examples finish up the paper to show the creation of the Security Building Blocks and their interactions with other software components.", "abstracts": [ { "abstractType": "Regular", "content": "In todayâs software development process, security related design decisions are rarely made early in the overall process. Even if security is considered early, this means that in most cases a more-or-less encompassing security requirement analyses is made, Based on this analysis best-practices, ad-hocdesign decisions or individual expertise is used to integrate security during the development process or after weaknesses are found after the deployment. This paper introduces Security Building Block Models which are used to build security related components, namely Security Building Blocks. These Security Building Blocks represent concrete security solutions, so called Security Properties, introduced in other publications of the Sec Futur project. The goal of this approach is to provide already defined and tested security related software components, which can be used early in the overall development process, to support security-design-decision already while modeling the software-system. The paper shortly describes this new Security Engineering Process with its requirement analysis and definition of Security Properties and how the Security Building Block Model fits into this approach. Additionally the Security Building Block Model is presented in detail. All artifacts and relationships of the model are described. Short examples finish up the paper to show the creation of the Security Building Blocks and their interactions with other software components.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In todayâs software development process, security related design decisions are rarely made early in the overall process. Even if security is considered early, this means that in most cases a more-or-less encompassing security requirement analyses is made, Based on this analysis best-practices, ad-hocdesign decisions or individual expertise is used to integrate security during the development process or after weaknesses are found after the deployment. This paper introduces Security Building Block Models which are used to build security related components, namely Security Building Blocks. These Security Building Blocks represent concrete security solutions, so called Security Properties, introduced in other publications of the Sec Futur project. The goal of this approach is to provide already defined and tested security related software components, which can be used early in the overall development process, to support security-design-decision already while modeling the software-system. The paper shortly describes this new Security Engineering Process with its requirement analysis and definition of Security Properties and how the Security Building Block Model fits into this approach. Additionally the Security Building Block Model is presented in detail. All artifacts and relationships of the model are described. Short examples finish up the paper to show the creation of the Security Building Blocks and their interactions with other software components.", "fno": "4938a132", "keywords": [ "Modelling", "Security Engineering", "Security Models And Meta Models", "Security Building Blocks", "Integration Of Components Component" ], "authors": [ { "affiliation": null, "fullName": "Andre Rein", "givenName": "Andre", "surname": "Rein", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Carsten Rudolph", "givenName": "Carsten", "surname": "Rudolph", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jose Fran. Ruiz", "givenName": "Jose Fran.", "surname": "Ruiz", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Marcos Arjona", "givenName": "Marcos", "surname": "Arjona", "__typename": "ArticleAuthorType" } ], "idPrefix": "biomedcom", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-12-01T00:00:00", "pubType": "proceedings", "pages": "132-139", "year": "2012", "issn": null, "isbn": "978-1-4673-5495-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4938a127", "articleId": "12OmNzZ5on0", "__typename": "AdjacentArticleType" }, "next": { "fno": "4938a140", "articleId": "12OmNzBwGpn", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cso/2009/3605/2/3605c989", "title": "Modeling and Optimization of Hull Building Process Based on P-TPN", "doi": null, "abstractUrl": "/proceedings-article/cso/2009/3605c989/12OmNBBhN80", "parentPublication": { "id": "proceedings/cso/2009/3605/2", "title": "2009 International Joint Conference on Computational Sciences and Optimization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cit/2017/0958/0/0958a332", "title": "Security Assessment of a Distributed, Modbus-Based Building Automation System", "doi": null, "abstractUrl": "/proceedings-article/cit/2017/0958a332/12OmNvzJG65", "parentPublication": { "id": "proceedings/cit/2017/0958/0", "title": "2017 IEEE International Conference on Computer and Information Technology (CIT)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dac/1988/0864/0/00014816", "title": "Connectivity biased channel construction and ordering for building-block layout", "doi": null, "abstractUrl": "/proceedings-article/dac/1988/00014816/12OmNx7G65z", "parentPublication": { "id": "proceedings/dac/1988/0864/0", "title": "Design Automation Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mmsd/1996/7511/0/75110041", "title": "Multimedia Building Blocks for Distributed Applications", "doi": null, "abstractUrl": "/proceedings-article/mmsd/1996/75110041/12OmNxWcHfs", "parentPublication": { "id": "proceedings/mmsd/1996/7511/0", "title": "Multimedia Software Development, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icip/1997/8183/2/81832152", "title": "Block Matching Motion Estimation Using Block Integration Based on Reliability Metric", "doi": null, "abstractUrl": "/proceedings-article/icip/1997/81832152/12OmNy3RRN5", "parentPublication": { "id": "proceedings/icip/1997/8183/2", "title": "Proceedings of International Conference on Image Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/biomedcom/2012/4938/0/4938a113", "title": "Security Engineering and Modelling of Set-Top Boxes", "doi": null, "abstractUrl": "/proceedings-article/biomedcom/2012/4938a113/12OmNySosN2", "parentPublication": { "id": "proceedings/biomedcom/2012/4938/0", "title": "ASE/IEEE International Conference on BioMedical Computing (BioMedCom)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icm/2011/4522/2/4522b034", "title": "Component Warehouse Driven by Workflow Based on Building Block", "doi": null, "abstractUrl": "/proceedings-article/icm/2011/4522b034/12OmNyaoDx2", "parentPublication": { "id": "proceedings/icm/2011/4522/2", "title": "Information Technology, Computer Engineering and Management Sciences, International Conference of", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscc/1995/7075/0/70750157", "title": "Security associations: building blocks for secure communications", "doi": null, "abstractUrl": "/proceedings-article/iscc/1995/70750157/12OmNyo1nRV", "parentPublication": { "id": "proceedings/iscc/1995/7075/0", "title": "Proceedings IEEE Symposium on Computers and Communications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/auic/2000/0515/0/05150141", "title": "Programming without a Computer: A New Interface for Children under Eight", "doi": null, "abstractUrl": "/proceedings-article/auic/2000/05150141/12OmNzuZUDW", "parentPublication": { "id": "proceedings/auic/2000/0515/0", "title": "Australasian User Interface Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tc/1969/08/01671345", "title": "Digital Adaptive-Element Building Blocks for MOS Large-Scale Integration", "doi": null, "abstractUrl": "/journal/tc/1969/08/01671345/13rRUNvya7V", "parentPublication": { "id": "trans/tc", "title": "IEEE Transactions on Computers", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzVGcIy", "title": "Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "1997", "__typename": "ProceedingType" }, "article": { "id": "12OmNzSh18X", "doi": "10.1109/CVPR.1997.609463", "title": "Hyper-patches for 3D model acquisition and tracking", "normalizedTitle": "Hyper-patches for 3D model acquisition and tracking", "abstract": "Automatic 3D model acquisition and 3D tracking of simple objects under motion using a single camera is often difficult due to the sparsity of information from which to establish the model. We have developed an automatic scheme that first computes a simple pointalistic Euclidean model of the object and then enriches this model using hyper-patches. These hyper-patches contain information on both the orientation and intensity pattern variation of roughly planar patches on an object. This information allows both the spatial and intensity distortions of the projected patch to be modelled accurately under 3D object motion. We show that hyper-patches not only can be computed automatically during model acquisition from a monocular image sequence, but that they are also extremely appropriate for the task of visual tracking.", "abstracts": [ { "abstractType": "Regular", "content": "Automatic 3D model acquisition and 3D tracking of simple objects under motion using a single camera is often difficult due to the sparsity of information from which to establish the model. We have developed an automatic scheme that first computes a simple pointalistic Euclidean model of the object and then enriches this model using hyper-patches. These hyper-patches contain information on both the orientation and intensity pattern variation of roughly planar patches on an object. This information allows both the spatial and intensity distortions of the projected patch to be modelled accurately under 3D object motion. We show that hyper-patches not only can be computed automatically during model acquisition from a monocular image sequence, but that they are also extremely appropriate for the task of visual tracking.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Automatic 3D model acquisition and 3D tracking of simple objects under motion using a single camera is often difficult due to the sparsity of information from which to establish the model. We have developed an automatic scheme that first computes a simple pointalistic Euclidean model of the object and then enriches this model using hyper-patches. These hyper-patches contain information on both the orientation and intensity pattern variation of roughly planar patches on an object. This information allows both the spatial and intensity distortions of the projected patch to be modelled accurately under 3D object motion. We show that hyper-patches not only can be computed automatically during model acquisition from a monocular image sequence, but that they are also extremely appropriate for the task of visual tracking.", "fno": "78221074", "keywords": [ "Solid Modelling Hyper Patches 3 D Model Acquisition 3 D Model Tracking Pointalistic Euclidean Model Orientation Intensity Pattern Variation 3 D Object Motion Model Acquisition Monocular Image Sequence Visual Tracking" ], "authors": [ { "affiliation": "Kansai Res. Labs., Toshiba Corp., Kobe, Japan", "fullName": "C.S. Wiles", "givenName": "C.S.", "surname": "Wiles", "__typename": "ArticleAuthorType" }, { "affiliation": "Kansai Res. Labs., Toshiba Corp., Kobe, Japan", "fullName": "A. Maki", "givenName": "A.", "surname": "Maki", "__typename": "ArticleAuthorType" }, { "affiliation": "Kansai Res. Labs., Toshiba Corp., Kobe, Japan", "fullName": "N. Matsuda", "givenName": "N.", "surname": "Matsuda", "__typename": "ArticleAuthorType" }, { "affiliation": "Kansai Res. Labs., Toshiba Corp., Kobe, Japan", "fullName": "M. Watanabe", "givenName": "M.", "surname": "Watanabe", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1997-06-01T00:00:00", "pubType": "proceedings", "pages": "1074", "year": "1997", "issn": "1063-6919", "isbn": "0-8186-7822-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "78221067", "articleId": "12OmNzcPACL", "__typename": "AdjacentArticleType" }, "next": { "fno": "78221081", "articleId": "12OmNyuPLbW", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxQOjzD", "title": "Visualization Conference, IEEE", "acronym": "ieee-vis", "groupId": "1000796", "volume": "0", "displayVolume": "0", "year": "1997", "__typename": "ProceedingType" }, "article": { "id": "12OmNBUS74c", "doi": "10.1109/VISUAL.1997.663882", "title": "An anti-aliasing technique for splatting", "normalizedTitle": "An anti-aliasing technique for splatting", "abstract": "Splatting is a popular direct volume rendering algorithm. However, the algorithm does not correctly render cases where the volume sampling rate is higher than the image sampling rate (e.g. more than one voxel maps into a pixel). This situation arises with orthographic projections of high-resolution volumes, as well as with perspective projections of volumes of any resolution. The result is potentially severe spatial and temporal aliasing artifacts. Some volume ray-casting algorithms avoid these artifacts by employing reconstruction kernels which vary in width as the rays diverge. Unlike ray-casting algorithms, existing splatting algorithms do not have an equivalent mechanism for avoiding these artifacts. The authors propose such a mechanism, which delivers high-quality splatted images and has the potential for a very efficient hardware implementation.", "abstracts": [ { "abstractType": "Regular", "content": "Splatting is a popular direct volume rendering algorithm. However, the algorithm does not correctly render cases where the volume sampling rate is higher than the image sampling rate (e.g. more than one voxel maps into a pixel). This situation arises with orthographic projections of high-resolution volumes, as well as with perspective projections of volumes of any resolution. The result is potentially severe spatial and temporal aliasing artifacts. Some volume ray-casting algorithms avoid these artifacts by employing reconstruction kernels which vary in width as the rays diverge. Unlike ray-casting algorithms, existing splatting algorithms do not have an equivalent mechanism for avoiding these artifacts. The authors propose such a mechanism, which delivers high-quality splatted images and has the potential for a very efficient hardware implementation.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Splatting is a popular direct volume rendering algorithm. However, the algorithm does not correctly render cases where the volume sampling rate is higher than the image sampling rate (e.g. more than one voxel maps into a pixel). This situation arises with orthographic projections of high-resolution volumes, as well as with perspective projections of volumes of any resolution. The result is potentially severe spatial and temporal aliasing artifacts. Some volume ray-casting algorithms avoid these artifacts by employing reconstruction kernels which vary in width as the rays diverge. Unlike ray-casting algorithms, existing splatting algorithms do not have an equivalent mechanism for avoiding these artifacts. The authors propose such a mechanism, which delivers high-quality splatted images and has the potential for a very efficient hardware implementation.", "fno": "82620197", "keywords": [ "Rendering Computer Graphics Splatting Direct Volume Rendering Algorithm Volume Sampling Rate Image Sampling Rate Orthographic Projections High Resolution Volumes Perspective Projections Spatial Aliasing Artifacts Temporal Aliasing Artifacts High Quality Splatted Images Efficient Hardware Implementation Anti Aliasing Technique" ], "authors": [ { "affiliation": "Adv. Comput. Center for the Arts & Design, Ohio State Univ., Columbus, OH, USA", "fullName": "J.E. Swan, II", "givenName": "J.E.", "surname": "Swan, II", "__typename": "ArticleAuthorType" }, { "affiliation": "Adv. Comput. Center for the Arts & Design, Ohio State Univ., Columbus, OH, USA", "fullName": "K. Mueller", "givenName": "K.", "surname": "Mueller", "__typename": "ArticleAuthorType" }, { "affiliation": "Adv. Comput. Center for the Arts & Design, Ohio State Univ., Columbus, OH, USA", "fullName": "T. Moller", "givenName": "T.", "surname": "Moller", "__typename": "ArticleAuthorType" }, { "affiliation": "Adv. Comput. Center for the Arts & Design, Ohio State Univ., Columbus, OH, USA", "fullName": "N. Shareel", "givenName": "N.", "surname": "Shareel", "__typename": "ArticleAuthorType" }, { "affiliation": "Adv. Comput. Center for the Arts & Design, Ohio State Univ., Columbus, OH, USA", "fullName": "R. Crawfis", "givenName": "R.", "surname": "Crawfis", "__typename": "ArticleAuthorType" }, { "affiliation": "Adv. Comput. Center for the Arts & Design, Ohio State Univ., Columbus, OH, USA", "fullName": "R. Yagel", "givenName": "R.", "surname": "Yagel", "__typename": "ArticleAuthorType" } ], "idPrefix": "ieee-vis", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1997-10-01T00:00:00", "pubType": "proceedings", "pages": "197", "year": "1997", "issn": "1070-2385", "isbn": "0-8186-8262-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "82620191", "articleId": "12OmNzVoBzB", "__typename": "AdjacentArticleType" }, "next": { "fno": "82620205", "articleId": "12OmNzQhP7p", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyoiYVr", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNxuo0id", "doi": "10.1109/CVPR.2017.97", "title": "Shading Annotations in the Wild", "normalizedTitle": "Shading Annotations in the Wild", "abstract": "Understanding shading effects in images is critical for a variety of vision and graphics problems, including intrinsic image decomposition, shadow removal, image relighting, and inverse rendering. As is the case with other vision tasks, machine learning is a promising approach to understanding shading - but there is little ground truth shading data available for real-world images. We introduce Shading Annotations in the Wild (SAW), a new large-scale, public dataset of shading annotations in indoor scenes, comprised of multiple forms of shading judgments obtained via crowdsourcing, along with shading annotations automatically generated from RGB-D imagery. We use this data to train a convolutional neural network to predict per-pixel shading information in an image. We demonstrate the value of our data and network in an application to intrinsic images, where we can reduce decomposition artifacts produced by existing algorithms. Our database is available at http://opensurfaces.cs.cornell.edu/saw.", "abstracts": [ { "abstractType": "Regular", "content": "Understanding shading effects in images is critical for a variety of vision and graphics problems, including intrinsic image decomposition, shadow removal, image relighting, and inverse rendering. As is the case with other vision tasks, machine learning is a promising approach to understanding shading - but there is little ground truth shading data available for real-world images. We introduce Shading Annotations in the Wild (SAW), a new large-scale, public dataset of shading annotations in indoor scenes, comprised of multiple forms of shading judgments obtained via crowdsourcing, along with shading annotations automatically generated from RGB-D imagery. We use this data to train a convolutional neural network to predict per-pixel shading information in an image. We demonstrate the value of our data and network in an application to intrinsic images, where we can reduce decomposition artifacts produced by existing algorithms. Our database is available at http://opensurfaces.cs.cornell.edu/saw.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Understanding shading effects in images is critical for a variety of vision and graphics problems, including intrinsic image decomposition, shadow removal, image relighting, and inverse rendering. As is the case with other vision tasks, machine learning is a promising approach to understanding shading - but there is little ground truth shading data available for real-world images. We introduce Shading Annotations in the Wild (SAW), a new large-scale, public dataset of shading annotations in indoor scenes, comprised of multiple forms of shading judgments obtained via crowdsourcing, along with shading annotations automatically generated from RGB-D imagery. We use this data to train a convolutional neural network to predict per-pixel shading information in an image. We demonstrate the value of our data and network in an application to intrinsic images, where we can reduce decomposition artifacts produced by existing algorithms. Our database is available at http://opensurfaces.cs.cornell.edu/saw.", "fno": "0457a850", "keywords": [ "Computer Vision", "Image Annotation", "Image Colour Analysis", "Learning Artificial Intelligence", "Neural Nets", "Graphics Problems", "Intrinsic Image Decomposition", "Shadow Removal", "Image Relighting", "Inverse Rendering", "Vision Tasks", "Machine Learning", "Shading Annotations In The Wild", "Indoor Scenes", "RGB D Imagery", "Convolutional Neural Network Training", "Per Pixel Shading Information Prediction", "Decomposition Artifacts Reduction", "Lighting", "Shape", "Image Decomposition", "Surface Treatment", "Image Color Analysis", "Surface Acoustic Waves" ], "authors": [ { "affiliation": null, "fullName": "Balazs Kovacs", "givenName": "Balazs", "surname": "Kovacs", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Sean Bell", "givenName": "Sean", "surname": "Bell", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Noah Snavely", "givenName": "Noah", "surname": "Snavely", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Kavita Bala", "givenName": "Kavita", "surname": "Bala", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-07-01T00:00:00", "pubType": "proceedings", "pages": "850-859", "year": "2017", "issn": "1063-6919", "isbn": "978-1-5386-0457-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0457a840", "articleId": "12OmNvlPkCJ", "__typename": "AdjacentArticleType" }, "next": { "fno": "0457a860", "articleId": "12OmNvAiS92", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/acssc/1988/9999/1/00754000", "title": "Shape From Shading In Sar Imagery: Experimental Results", "doi": null, "abstractUrl": "/proceedings-article/acssc/1988/00754000/12OmNAoUTq2", "parentPublication": { "id": "proceedings/acssc/1988/9999/1", "title": "Twenty-Second Asilomar Conference on Signals, Systems and Computers", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ssst/1990/2038/0/00138173", "title": "CAD tools for the design and analysis of surface acoustic wave chirp devices", "doi": null, "abstractUrl": "/proceedings-article/ssst/1990/00138173/12OmNCcbE7B", "parentPublication": { "id": "proceedings/ssst/1990/2038/0", "title": "Proceedings The Twenty-Second Southeastern Symposium on System Theory", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/haptics/2010/6821/0/05444639", "title": "Tactile mouse using friction control", "doi": null, "abstractUrl": "/proceedings-article/haptics/2010/05444639/12OmNwF0BXG", "parentPublication": { "id": "proceedings/haptics/2010/6821/0", "title": "2010 IEEE Haptics Symposium (Formerly known as Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/is3c/2018/7036/0/703600a392", "title": "Study of In Vitro Diagnostic Preconcentration Technique via Surface Acoustic Wave Modified with the 54 MHz Communication Chip", "doi": null, "abstractUrl": "/proceedings-article/is3c/2018/703600a392/17QjJfqxwoU", "parentPublication": { "id": "proceedings/is3c/2018/7036/0", "title": "2018 International Symposium on Computer, Consumer and Control (IS3C)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900a312", "title": "HSI-Guided Intrinsic Image Decomposition for Outdoor Scenes", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900a312/1G56nWipNPa", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ccpqt/2022/7020/0/702000a328", "title": "Research on Detection Technology of Bio-electronic Nose Based on SAW Devices", "doi": null, "abstractUrl": "/proceedings-article/ccpqt/2022/702000a328/1Iiu25VEU80", "parentPublication": { "id": "proceedings/ccpqt/2022/7020/0", "title": "2022 International Conference on Computing, Communication, Perception and Quantum Technology (CCPQT)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2019/9552/0/955200a175", "title": "Towards High-Quality Intrinsic Images in the Wild", "doi": null, "abstractUrl": "/proceedings-article/icme/2019/955200a175/1cdOQdLofSw", "parentPublication": { "id": "proceedings/icme/2019/9552/0", "title": "2019 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300h819", "title": "GLoSH: Global-Local Spherical Harmonics for Intrinsic Image Decomposition", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300h819/1hQqy771H9u", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccis/2019/6297/0/09151864", "title": "Analysis of Bioproteins by Propagation Signal Based on SAW Device", "doi": null, "abstractUrl": "/proceedings-article/iccis/2019/09151864/1lRlKMvW4fK", "parentPublication": { "id": "proceedings/iccis/2019/6297/0", "title": "2019 4th International Conference on Communication and Information Systems (ICCIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/12/09625763", "title": "Unsupervised Intrinsic Image Decomposition Using Internal Self-Similarity Cues", "doi": null, "abstractUrl": "/journal/tp/2022/12/09625763/1yLTnG9Uisw", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAZOJTa", "title": "Point-Based Graphics 2005", "acronym": "pbg", "groupId": "1002154", "volume": "0", "displayVolume": "0", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "12OmNzdoMWS", "doi": "10.1109/PBG.2005.194059", "title": "High-quality surface splatting on today's GPUs", "normalizedTitle": "High-quality surface splatting on today's GPUs", "abstract": "Point-based geometries evolved into a valuable alternative to surface representations based on polygonal meshes, because of their conceptual simplicity and superior flexibility. Elliptical surface splats were shown to allow for high-quality anti-aliased rendering by sophisticated EWA filtering. Since the publication of the original software-based EWA splatting, several authors tried to map this technique to the GPU in order to exploit hardware acceleration. Due to the lacking support for splat primitives, these methods always have to find a trade-off between rendering quality and rendering performance. In this paper, we discuss the capabilities of today's GPUs for hardware-accelerated surface splatting. We present an approach that achieves a quality comparable to the original EWA splatting at a rate of more than 20M elliptical splats per second. In contrast to previous GPU renderers, our method provides per-pixel Phong shading even for dynamically changing geometries and high-quality anti-aliasing by employing a screen-space pre-filter in addition to the object-space reconstruction filter. The use of deferred shading techniques effectively avoids unnecessary shader computations and additionally provides a clear separation between the rasterization and the shading of elliptical splats, which considerably simplifies the development of custom shaders. We demonstrate quality, efficiency, and flexibility of our approach by showing several shaders on a range of models.", "abstracts": [ { "abstractType": "Regular", "content": "Point-based geometries evolved into a valuable alternative to surface representations based on polygonal meshes, because of their conceptual simplicity and superior flexibility. Elliptical surface splats were shown to allow for high-quality anti-aliased rendering by sophisticated EWA filtering. Since the publication of the original software-based EWA splatting, several authors tried to map this technique to the GPU in order to exploit hardware acceleration. Due to the lacking support for splat primitives, these methods always have to find a trade-off between rendering quality and rendering performance. In this paper, we discuss the capabilities of today's GPUs for hardware-accelerated surface splatting. We present an approach that achieves a quality comparable to the original EWA splatting at a rate of more than 20M elliptical splats per second. In contrast to previous GPU renderers, our method provides per-pixel Phong shading even for dynamically changing geometries and high-quality anti-aliasing by employing a screen-space pre-filter in addition to the object-space reconstruction filter. The use of deferred shading techniques effectively avoids unnecessary shader computations and additionally provides a clear separation between the rasterization and the shading of elliptical splats, which considerably simplifies the development of custom shaders. We demonstrate quality, efficiency, and flexibility of our approach by showing several shaders on a range of models.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Point-based geometries evolved into a valuable alternative to surface representations based on polygonal meshes, because of their conceptual simplicity and superior flexibility. Elliptical surface splats were shown to allow for high-quality anti-aliased rendering by sophisticated EWA filtering. Since the publication of the original software-based EWA splatting, several authors tried to map this technique to the GPU in order to exploit hardware acceleration. Due to the lacking support for splat primitives, these methods always have to find a trade-off between rendering quality and rendering performance. In this paper, we discuss the capabilities of today's GPUs for hardware-accelerated surface splatting. We present an approach that achieves a quality comparable to the original EWA splatting at a rate of more than 20M elliptical splats per second. In contrast to previous GPU renderers, our method provides per-pixel Phong shading even for dynamically changing geometries and high-quality anti-aliasing by employing a screen-space pre-filter in addition to the object-space reconstruction filter. The use of deferred shading techniques effectively avoids unnecessary shader computations and additionally provides a clear separation between the rasterization and the shading of elliptical splats, which considerably simplifies the development of custom shaders. We demonstrate quality, efficiency, and flexibility of our approach by showing several shaders on a range of models.", "fno": "01500313", "keywords": [ "Rendering Computer Graphics", "Computational Geometry", "Surface Fitting", "Mesh Generation", "Computer Graphic Equipment", "High Quality Surface Splatting", "GPU", "Point Based Geometry", "Surface Representation", "Polygonal Mesh", "High Quality Anti Aliased Rendering", "Sophisticated EWA Filtering", "Software Based EWA Splatting", "Hardware Acceleration", "Per Pixel Phong Shading", "Object Space Reconstruction Filter", "Surface Reconstruction", "Rendering Computer Graphics", "Filtering", "Hardware", "Computer Graphics", "Computational Geometry", "Sampling Methods", "Acceleration", "Information Geometry", "Piecewise Linear Techniques" ], "authors": [ { "affiliation": "Comput. Graphics Group, RWTH Aachen Tech. Univ., Germany", "fullName": "M. Botsch", "givenName": "M.", "surname": "Botsch", "__typename": "ArticleAuthorType" }, { "affiliation": "Comput. Graphics Group, RWTH Aachen Tech. Univ., Germany", "fullName": "A. Hornung", "givenName": "A.", "surname": "Hornung", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "M. Zwicker", "givenName": "M.", "surname": "Zwicker", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "L. Kobbelt", "givenName": "L.", "surname": "Kobbelt", "__typename": "ArticleAuthorType" } ], "idPrefix": "pbg", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2005-03-01T00:00:00", "pubType": "proceedings", "pages": "17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118,119,120,121,122,123,124,125,126,127,128,129,130,131,132,133,134,135,136,137,138,139,140,141", "year": "2005", "issn": "1511-7813", "isbn": "3-905673-20-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "01500312", "articleId": "12OmNxaeu1E", "__typename": "AdjacentArticleType" }, "next": { "fno": "01500314", "articleId": "12OmNqNos8r", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ieee-vis/1999/5897/0/58970061", "title": "Splatting Without The Blur", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/1999/58970061/12OmNAoDhUR", "parentPublication": { "id": "proceedings/ieee-vis/1999/5897/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/visual/1996/864/0/00567608", "title": "Fast perspective volume rendering with splatting by utilizing a ray-driven approach", "doi": null, "abstractUrl": "/proceedings-article/visual/1996/00567608/12OmNC4eSCW", "parentPublication": { "id": "proceedings/visual/1996/864/0", "title": "Proceedings of Seventh Annual IEEE Visualization '96", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pbg/2005/20/0/01500314", "title": "Progressive splatting", "doi": null, "abstractUrl": "/proceedings-article/pbg/2005/01500314/12OmNqNos8r", "parentPublication": { "id": "proceedings/pbg/2005/20/0", "title": "Point-Based Graphics 2005", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/27660035", "title": "High Performance Volume Splatting for Visualization of Neurovascular Data", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660035/12OmNrYlmLf", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vg/2005/26/0/01500542", "title": "GPU accelerated image aligned splatting", "doi": null, "abstractUrl": "/proceedings-article/vg/2005/01500542/12OmNxX3uw9", "parentPublication": { "id": "proceedings/vg/2005/26/0", "title": "Volume Graphics 2005", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2001/7200/0/7200zwicker", "title": "EWA Volume Splatting", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2001/7200zwicker/12OmNxwWoHl", "parentPublication": { "id": "proceedings/ieee-vis/2001/7200/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/2003/1946/0/19460196", "title": "Fast Depth of Field Rendering with Surface Splatting", "doi": null, "abstractUrl": "/proceedings-article/cgi/2003/19460196/12OmNyKrHh5", "parentPublication": { "id": "proceedings/cgi/2003/1946/0", "title": "Computer Graphics International Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/01532805", "title": "High performance volume splatting for visualization of neurovascular data", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/01532805/12OmNyQGS8K", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2004/8788/0/87880067", "title": "Hardware-Accelerated Adaptive EWA Volume Splatting", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2004/87880067/12OmNzX6cvd", "parentPublication": { "id": "proceedings/ieee-vis/2004/8788/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2002/03/v0223", "title": "EWA Splatting", "doi": null, "abstractUrl": "/journal/tg/2002/03/v0223/13rRUxNEqPB", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyYm2wA", "title": "Asia-Pacific Symposium on Visualisation 2007", "acronym": "apvis", "groupId": "1001657", "volume": "0", "displayVolume": "0", "year": "2007", "__typename": "ProceedingType" }, "article": { "id": "12OmNARRYyX", "doi": "10.1109/APVIS.2007.329285", "title": "Adaptive sampling in three dimensions for volume rendering on GPUs", "normalizedTitle": "Adaptive sampling in three dimensions for volume rendering on GPUs", "abstract": "Direct volume rendering of large volumetric data sets on programmable graphics hardware is often limited by the amount of available graphics memory and the bandwidth from main memory to graphics memory. Therefore, several approaches to volume rendering from compact representations of volumetric data have been published that avoid most of the data transfer between main memory and the graphics programming unit (GPU) at the cost of additional data decompression by the GPU. To reduce this performance cost, adaptive sampling techniques were proposed; which are, however, usually restricted to the sampling in view direction. In this work, we present a GPU-based volume rendering algorithm with adaptive sampling in all three spatial directions; i.e., not only in view direction but also in the two perpendicular directions of the image plane. This approach allows us to reduce the number of samples dramatically without compromising image quality; thus, it is particularly well suited for many compressed representations of volumetric data that require a computational expensive GPU-based sampling of data", "abstracts": [ { "abstractType": "Regular", "content": "Direct volume rendering of large volumetric data sets on programmable graphics hardware is often limited by the amount of available graphics memory and the bandwidth from main memory to graphics memory. Therefore, several approaches to volume rendering from compact representations of volumetric data have been published that avoid most of the data transfer between main memory and the graphics programming unit (GPU) at the cost of additional data decompression by the GPU. To reduce this performance cost, adaptive sampling techniques were proposed; which are, however, usually restricted to the sampling in view direction. In this work, we present a GPU-based volume rendering algorithm with adaptive sampling in all three spatial directions; i.e., not only in view direction but also in the two perpendicular directions of the image plane. This approach allows us to reduce the number of samples dramatically without compromising image quality; thus, it is particularly well suited for many compressed representations of volumetric data that require a computational expensive GPU-based sampling of data", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Direct volume rendering of large volumetric data sets on programmable graphics hardware is often limited by the amount of available graphics memory and the bandwidth from main memory to graphics memory. Therefore, several approaches to volume rendering from compact representations of volumetric data have been published that avoid most of the data transfer between main memory and the graphics programming unit (GPU) at the cost of additional data decompression by the GPU. To reduce this performance cost, adaptive sampling techniques were proposed; which are, however, usually restricted to the sampling in view direction. In this work, we present a GPU-based volume rendering algorithm with adaptive sampling in all three spatial directions; i.e., not only in view direction but also in the two perpendicular directions of the image plane. This approach allows us to reduce the number of samples dramatically without compromising image quality; thus, it is particularly well suited for many compressed representations of volumetric data that require a computational expensive GPU-based sampling of data", "fno": "04126228", "keywords": [ "Volume Visualization", "GPU Based Volume Rendering Algorithm", "Large Volumetric Data Set Direct Volume Rendering", "Programmable Graphics Hardware", "Graphics Memory", "Main Memory", "Data Transfer", "Graphics Programming Unit", "Data Decompression", "Image Quality", "Volumetric Data Compressed Representation", "Adaptive GPU Based Data Sampling", "Scientific Visualisation" ], "authors": [ { "affiliation": "Visualization&Interactive Syst. Group, Stuttgart Univ.", "fullName": "M. Kraus", "givenName": "M.", "surname": "Kraus", "__typename": "ArticleAuthorType" }, { "affiliation": "Visualization&Interactive Syst. Group, Stuttgart Univ.", "fullName": "M. Strengert", "givenName": "M.", "surname": "Strengert", "__typename": "ArticleAuthorType" }, { "affiliation": "Visualization&Interactive Syst. Group, Stuttgart Univ.", "fullName": "T. Klein", "givenName": "T.", "surname": "Klein", "__typename": "ArticleAuthorType" }, { "affiliation": "Visualization&Interactive Syst. Group, Stuttgart Univ.", "fullName": "T. Ertl", "givenName": "T.", "surname": "Ertl", "__typename": "ArticleAuthorType" } ], "idPrefix": "apvis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2007-02-01T00:00:00", "pubType": "proceedings", "pages": "113-120", "year": "2007", "issn": null, "isbn": "1-4244-0808-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "04126227", "articleId": "12OmNBOCWfN", "__typename": "AdjacentArticleType" }, "next": { "fno": "04126229", "articleId": "12OmNzaQogz", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ieee-vis/2003/2030/0/20300039", "title": "Compression Domain Volume Rendering", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2003/20300039/12OmNBWi6GJ", "parentPublication": { "id": "proceedings/ieee-vis/2003/2030/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2004/8788/0/87880027", "title": "Projecting Tetrahedra without Rendering Artifacts", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2004/87880027/12OmNBbsidz", "parentPublication": { "id": "proceedings/ieee-vis/2004/8788/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2003/2030/0/20300038", "title": "Acceleration Techniques for GPU-based Volume Rendering", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2003/20300038/12OmNC2xhD8", "parentPublication": { "id": "proceedings/ieee-vis/2003/2030/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/esiat/2009/3682/2/3682b575", "title": "Rapid Texture-based Volume Rendering", "doi": null, "abstractUrl": "/proceedings-article/esiat/2009/3682b575/12OmNx7G5VW", "parentPublication": { "id": "esiat/2009/3682/2", "title": "Environmental Science and Information Application Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/27660029", "title": "Exploiting Frame-to-Frame Coherence for Accelerating High-Quality Volume Raycasting on Graphics Hardware", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660029/12OmNzt0IIb", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/10/07378994", "title": "Compressed Coverage Masks for Path Rendering on Mobile GPUs", "doi": null, "abstractUrl": "/journal/tg/2016/10/07378994/13rRUwIF6l9", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2006/05/v1345", "title": "A Generic and Scalable Pipeline for GPU Tetrahedral Grid Rendering", "doi": null, "abstractUrl": "/journal/tg/2006/05/v1345/13rRUwgQpDj", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2009/06/ttg2009061579", "title": "Interactive Volume Rendering of Functional Representations in Quantum Chemistry", "doi": null, "abstractUrl": "/journal/tg/2009/06/ttg2009061579/13rRUwgQpDq", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/06/v1600", "title": "Transform Coding for Hardware-accelerated Volume Rendering", "doi": null, "abstractUrl": "/journal/tg/2007/06/v1600/13rRUyeTVhV", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hcs/2022/6028/0/09895607", "title": "Accelerating Graphic Rendering on Programmable RISC-V GPUs", "doi": null, "abstractUrl": "/proceedings-article/hcs/2022/09895607/1GZiJcJFUk0", "parentPublication": { "id": "proceedings/hcs/2022/6028/0", "title": "2022 IEEE Hot Chips 34 Symposium (HCS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAYoKmj", "title": "2012 International Conference on Virtual Reality and Visualization", "acronym": "icvrv", "groupId": "1800579", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNBaBuS8", "doi": "10.1109/ICVRV.2012.8", "title": "GPU Based Compression and Rendering of Massive Aircraft CAD Models", "normalizedTitle": "GPU Based Compression and Rendering of Massive Aircraft CAD Models", "abstract": "The total size of massive aircraft CAD models is usually up to several GBs, which exceed not only the storage capacity of memory, but also the rendering ability of graphics card. In this paper, we present compression and rendering methods by exploring the up-to-date GPU techniques. To fit into the memory, vertex data are compressed from float to byte type with bounding box information and then decompressed with GPU. Index data are in short or byte type according to the vertex size, while normal data are deleted and generated by GPU while rendering. To render in real-time, vertex buffer object is exploited instead of traditional display list for efficiency and GPU occlusion query culls occluded parts to lower the rendering load. Furthermore, deliberately designed GPU shaders are applied to optimize the traditional rendering pipeline. The experiment results show by the GPU based methods, the compression rates get up to 5.3, massive CAD models such as the regional jet can be compressed within 178 MB and fit into memory of personal computers, and the rendering frame rates achieve up to 40 with cheap graphics card. It's proved that our method maximizes the GPU capabilities to accelerate the real-time rendering performance of massive aircraft CAD models.", "abstracts": [ { "abstractType": "Regular", "content": "The total size of massive aircraft CAD models is usually up to several GBs, which exceed not only the storage capacity of memory, but also the rendering ability of graphics card. In this paper, we present compression and rendering methods by exploring the up-to-date GPU techniques. To fit into the memory, vertex data are compressed from float to byte type with bounding box information and then decompressed with GPU. Index data are in short or byte type according to the vertex size, while normal data are deleted and generated by GPU while rendering. To render in real-time, vertex buffer object is exploited instead of traditional display list for efficiency and GPU occlusion query culls occluded parts to lower the rendering load. Furthermore, deliberately designed GPU shaders are applied to optimize the traditional rendering pipeline. The experiment results show by the GPU based methods, the compression rates get up to 5.3, massive CAD models such as the regional jet can be compressed within 178 MB and fit into memory of personal computers, and the rendering frame rates achieve up to 40 with cheap graphics card. It's proved that our method maximizes the GPU capabilities to accelerate the real-time rendering performance of massive aircraft CAD models.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The total size of massive aircraft CAD models is usually up to several GBs, which exceed not only the storage capacity of memory, but also the rendering ability of graphics card. In this paper, we present compression and rendering methods by exploring the up-to-date GPU techniques. To fit into the memory, vertex data are compressed from float to byte type with bounding box information and then decompressed with GPU. Index data are in short or byte type according to the vertex size, while normal data are deleted and generated by GPU while rendering. To render in real-time, vertex buffer object is exploited instead of traditional display list for efficiency and GPU occlusion query culls occluded parts to lower the rendering load. Furthermore, deliberately designed GPU shaders are applied to optimize the traditional rendering pipeline. The experiment results show by the GPU based methods, the compression rates get up to 5.3, massive CAD models such as the regional jet can be compressed within 178 MB and fit into memory of personal computers, and the rendering frame rates achieve up to 40 with cheap graphics card. It's proved that our method maximizes the GPU capabilities to accelerate the real-time rendering performance of massive aircraft CAD models.", "fno": "4836a050", "keywords": [ "Rendering Computer Graphics", "Graphics Processing Units", "Solid Modeling", "Design Automation", "Atmospheric Modeling", "Real Time Systems", "Computational Modeling", "Real Time Rendering", "Massive Models", "GPU", "Model Compression" ], "authors": [ { "affiliation": null, "fullName": "Tan Dunming", "givenName": "Tan", "surname": "Dunming", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zhao Gang", "givenName": "Zhao", "surname": "Gang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yu Lu", "givenName": "Yu", "surname": "Lu", "__typename": "ArticleAuthorType" } ], "idPrefix": "icvrv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-09-01T00:00:00", "pubType": "proceedings", "pages": "50-55", "year": "2012", "issn": null, "isbn": "978-1-4673-5154-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4836a046", "articleId": "12OmNx5piSz", "__typename": "AdjacentArticleType" }, "next": { "fno": "4836a056", "articleId": "12OmNB836Kj", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icmew/2016/1552/0/07574730", "title": "A parallel volume rendering method for massive data", "doi": null, "abstractUrl": "/proceedings-article/icmew/2016/07574730/12OmNx0A7P1", "parentPublication": { "id": "proceedings/icmew/2016/1552/0", "title": "2016 IEEE International Conference on Multimedia & Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2014/4258/0/4258a335", "title": "Instanced Rendering of Massive CAD Models Using Shape Matching", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2014/4258a335/12OmNx5piSk", "parentPublication": { "id": "proceedings/sibgrapi/2014/4258/0", "title": "2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2017/2219/0/2219a031", "title": "CAD Shape Grammar: Procedural Generation for Massive CAD Model", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2017/2219a031/12OmNy6qfJ2", "parentPublication": { "id": "proceedings/sibgrapi/2017/2219/0", "title": "2017 30th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2004/8788/0/87880131", "title": "Quick-VDR: Interactive View-Dependent Rendering of Massive Models", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2004/87880131/12OmNyRPgrt", "parentPublication": { "id": "proceedings/ieee-vis/2004/8788/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cad-graphics/2013/2576/0/06815014", "title": "Real-Time Label Visualization in Massive CAD Models", "doi": null, "abstractUrl": "/proceedings-article/cad-graphics/2013/06815014/12OmNyTwRd8", "parentPublication": { "id": "proceedings/cad-graphics/2013/2576/0", "title": "2013 International Conference on Computer-Aided Design and Computer Graphics (CAD/Graphics)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cadgraphics/2011/4497/0/4497a025", "title": "Direct Rendering of Solid CAD Models on the GPU", "doi": null, "abstractUrl": "/proceedings-article/cadgraphics/2011/4497a025/12OmNzb7Zpm", "parentPublication": { "id": "proceedings/cadgraphics/2011/4497/0", "title": "Computer-Aided Design and Computer Graphics, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2014/03/ttg2014030481", "title": "T-ReX: Interactive Global Illumination of Massive Models on Heterogeneous Computing Resources", "doi": null, "abstractUrl": "/journal/tg/2014/03/ttg2014030481/13rRUxAAT7F", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dcve/2018/5132/0/08637114", "title": "Multi-user predictive rendering on remote multi-GPU clusters", "doi": null, "abstractUrl": "/proceedings-article/3dcve/2018/08637114/17D45Vu1TyF", "parentPublication": { "id": "proceedings/3dcve/2018/5132/0", "title": "2018 IEEE Fourth VR International Workshop on Collaborative Virtual Environments (3DCVE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2018/9264/0/926400a234", "title": "Hybrid Cloud Rendering System for Massive CAD Models", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2018/926400a234/17D45XDIXXh", "parentPublication": { "id": "proceedings/sibgrapi/2018/9264/0", "title": "2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sose/2019/1442/0/144200a251", "title": "Efficient Rendering of Large-Scale CAD Models on a GPU Virtualization Architecture with Model Geometry Metrics", "doi": null, "abstractUrl": "/proceedings-article/sose/2019/144200a251/19RSyjj8lna", "parentPublication": { "id": "proceedings/sose/2019/1442/0", "title": "2019 IEEE International Conference on Service-Oriented System Engineering (SOSE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAQJzKb", "title": "2015 IEEE Pacific Visualization Symposium (PacificVis)", "acronym": "pacificvis", "groupId": "1001657", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNwkzulc", "doi": "10.1109/PACIFICVIS.2015.7156372", "title": "Computation-to-core mapping strategies for iso-surface volume rendering on GPUs", "normalizedTitle": "Computation-to-core mapping strategies for iso-surface volume rendering on GPUs", "abstract": "Ray casting algorithm is a major component of the direct volume rendering, which exhibits inherent parallelism, making it suitable for graphics processing units (GPUs). However, blindly mapping the ray casting algorithm on a GPU's complex parallel architecture can result in a magnitude of performance loss. In this paper, a novel computation-to-core mapping strategy, called Warp Marching, for the texture-based iso-surface volume rendering is introduced. We evaluate and compare this new strategy with the most commonly used existing mapping strategy. Texture cache performance and load balancing are the two major evaluation factors since they have significant consequences on the overall rendering performance. Through a series of real-life data experiments, we conclude that the texture cache performances of these two computation-to-core mapping strategies are significantly affected by the viewing direction; and the Warp Marching performs better in balancing workloads among threads and concurrent hardware components of a GPU.", "abstracts": [ { "abstractType": "Regular", "content": "Ray casting algorithm is a major component of the direct volume rendering, which exhibits inherent parallelism, making it suitable for graphics processing units (GPUs). However, blindly mapping the ray casting algorithm on a GPU's complex parallel architecture can result in a magnitude of performance loss. In this paper, a novel computation-to-core mapping strategy, called Warp Marching, for the texture-based iso-surface volume rendering is introduced. We evaluate and compare this new strategy with the most commonly used existing mapping strategy. Texture cache performance and load balancing are the two major evaluation factors since they have significant consequences on the overall rendering performance. Through a series of real-life data experiments, we conclude that the texture cache performances of these two computation-to-core mapping strategies are significantly affected by the viewing direction; and the Warp Marching performs better in balancing workloads among threads and concurrent hardware components of a GPU.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Ray casting algorithm is a major component of the direct volume rendering, which exhibits inherent parallelism, making it suitable for graphics processing units (GPUs). However, blindly mapping the ray casting algorithm on a GPU's complex parallel architecture can result in a magnitude of performance loss. In this paper, a novel computation-to-core mapping strategy, called Warp Marching, for the texture-based iso-surface volume rendering is introduced. We evaluate and compare this new strategy with the most commonly used existing mapping strategy. Texture cache performance and load balancing are the two major evaluation factors since they have significant consequences on the overall rendering performance. Through a series of real-life data experiments, we conclude that the texture cache performances of these two computation-to-core mapping strategies are significantly affected by the viewing direction; and the Warp Marching performs better in balancing workloads among threads and concurrent hardware components of a GPU.", "fno": "07156372", "keywords": [ "Graphics Processing Units", "Standards", "Rendering Computer Graphics", "Instruction Sets", "Load Management", "Hardware", "Casting" ], "authors": [ { "affiliation": "Virginia Tech, USA", "fullName": "Junpeng Wang", "givenName": "Junpeng", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "NVIDIA, USA", "fullName": "Fei Yang", "givenName": "Fei", "surname": "Yang", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, USA", "fullName": "Yong Cao", "givenName": "Yong", "surname": "Cao", "__typename": "ArticleAuthorType" } ], "idPrefix": "pacificvis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-04-01T00:00:00", "pubType": "proceedings", "pages": "153-157", "year": "2015", "issn": null, "isbn": "978-1-4673-6879-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07156371", "articleId": "12OmNzJbR3z", "__typename": "AdjacentArticleType" }, "next": { "fno": "07156373", "articleId": "12OmNxEBzlW", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ieee-vis/2003/2030/0/20300038", "title": "Acceleration Techniques for GPU-based Volume Rendering", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2003/20300038/12OmNC2xhD8", "parentPublication": { "id": "proceedings/ieee-vis/2003/2030/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2001/7200/0/7200westermann", "title": "Accelerated Volume Ray-Casting using Texture Mapping", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2001/7200westermann/12OmNCbU30D", "parentPublication": { "id": "proceedings/ieee-vis/2001/7200/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2010/7846/0/05571370", "title": "Preserving Coherent Illumination in Style Transfer Functions for Volume Rendering", "doi": null, "abstractUrl": "/proceedings-article/iv/2010/05571370/12OmNwF0BUx", "parentPublication": { "id": "proceedings/iv/2010/7846/0", "title": "2010 14th International Conference Information Visualisation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ipps/1995/7074/0/70740700", "title": "Parallel implementation of volume rendering on Denali graphics systems", "doi": null, "abstractUrl": "/proceedings-article/ipps/1995/70740700/12OmNx0RIQu", "parentPublication": { "id": "proceedings/ipps/1995/7074/0", "title": "Proceedings of 9th International Parallel Processing Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/esiat/2009/3682/2/3682b575", "title": "Rapid Texture-based Volume Rendering", "doi": null, "abstractUrl": "/proceedings-article/esiat/2009/3682b575/12OmNx7G5VW", "parentPublication": { "id": "esiat/2009/3682/2", "title": "Environmental Science and Information Application Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ldav/2014/5215/0/07013200", "title": "Cache-aware sampling strategies for texture-based ray casting on GPU", "doi": null, "abstractUrl": "/proceedings-article/ldav/2014/07013200/12OmNxWcH5i", "parentPublication": { "id": "proceedings/ldav/2014/5215/0", "title": "2014 IEEE 4th Symposium on Large Data Analysis and Visualization (LDAV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/prs/1993/4920/0/00586079", "title": "Segmented ray casting for data parallel volume rendering", "doi": null, "abstractUrl": "/proceedings-article/prs/1993/00586079/12OmNybfr4E", "parentPublication": { "id": "proceedings/prs/1993/4920/0", "title": "Proceedings of 1993 IEEE Parallel Rendering Symposium", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/2001/1007/0/10070306", "title": "A Marching Voxels Method for Surface Rendering of Volume Data", "doi": null, "abstractUrl": "/proceedings-article/cgi/2001/10070306/12OmNzdoN6A", "parentPublication": { "id": "proceedings/cgi/2001/1007/0", "title": "Proceedings. Computer Graphics International 2001", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/10/07378994", "title": "Compressed Coverage Masks for Path Rendering on Mobile GPUs", "doi": null, "abstractUrl": "/journal/tg/2016/10/07378994/13rRUwIF6l9", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccgiv/2022/9250/0/925000a183", "title": "Ellipsoidal ray casting algorithm", "doi": null, "abstractUrl": "/proceedings-article/iccgiv/2022/925000a183/1LxfqGjszTi", "parentPublication": { "id": "proceedings/iccgiv/2022/9250/0", "title": "2022 2nd International Conference on Computer Graphics, Image and Virtualization (ICCGIV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1GZiD8ztM9a", "title": "2022 IEEE Hot Chips 34 Symposium (HCS)", "acronym": "hcs", "groupId": "1812447", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1GZiJcJFUk0", "doi": "10.1109/HCS55958.2022.9895607", "title": "Accelerating Graphic Rendering on Programmable RISC-V GPUs", "normalizedTitle": "Accelerating Graphic Rendering on Programmable RISC-V GPUs", "abstract": "Graphics rendering remains one of the most compute-intensive and memory-bound applications of GPUs and has been driving their push for performance and energy efficiency since its inception. Early GPU architectures focused only on accelerating graphics rendering and implemented dedicated a fixed-function rendering units. Today&#x2019;s GPUs have become more programmable to address the complexity and diversity of modern graphics workloads while still accelerating several components of the graphics pipeline in fixed-function hardware.Generalizing the GPU microarchitecture and implement some of its graphics hardware blocks in software can save area that can be used to expand the generic pipeline, especially in mobile systems-on-chips environments where power and area is scarce.In this work, we propose a RISC-V-based hybrid GPU architecture that accelerates the graphics pipeline without paying the cost of a full hardware graphics pipeline. We evaluated the design on an Altera Arria 10 FPGA running at 200 MHz.", "abstracts": [ { "abstractType": "Regular", "content": "Graphics rendering remains one of the most compute-intensive and memory-bound applications of GPUs and has been driving their push for performance and energy efficiency since its inception. Early GPU architectures focused only on accelerating graphics rendering and implemented dedicated a fixed-function rendering units. Today&#x2019;s GPUs have become more programmable to address the complexity and diversity of modern graphics workloads while still accelerating several components of the graphics pipeline in fixed-function hardware.Generalizing the GPU microarchitecture and implement some of its graphics hardware blocks in software can save area that can be used to expand the generic pipeline, especially in mobile systems-on-chips environments where power and area is scarce.In this work, we propose a RISC-V-based hybrid GPU architecture that accelerates the graphics pipeline without paying the cost of a full hardware graphics pipeline. We evaluated the design on an Altera Arria 10 FPGA running at 200 MHz.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Graphics rendering remains one of the most compute-intensive and memory-bound applications of GPUs and has been driving their push for performance and energy efficiency since its inception. Early GPU architectures focused only on accelerating graphics rendering and implemented dedicated a fixed-function rendering units. Today’s GPUs have become more programmable to address the complexity and diversity of modern graphics workloads while still accelerating several components of the graphics pipeline in fixed-function hardware.Generalizing the GPU microarchitecture and implement some of its graphics hardware blocks in software can save area that can be used to expand the generic pipeline, especially in mobile systems-on-chips environments where power and area is scarce.In this work, we propose a RISC-V-based hybrid GPU architecture that accelerates the graphics pipeline without paying the cost of a full hardware graphics pipeline. We evaluated the design on an Altera Arria 10 FPGA running at 200 MHz.", "fno": "09895607", "keywords": [ "Energy Conservation", "Graphics Processing Units", "Reduced Instruction Set Computing", "Rendering Computer Graphics", "System On Chip", "Graphics Rendering", "Compute Intensive Memory Bound Applications", "Energy Efficiency", "Fixed Function Rendering Units", "Modern Graphics Workloads", "Fixed Function Hardware", "GPU Microarchitecture", "Graphics Hardware Blocks", "RISC V Based Hybrid GPU Architecture", "Hardware Graphics Pipeline", "Programmable RISC V", "Memory Bound Applications", "Mobile System On Chip Environments", "Altera Arria 10 FPGA", "Frequency 200 0 M Hz", "Microarchitecture", "Pipelines", "Graphics Processing Units", "Computer Architecture", "Rendering Computer Graphics", "Hardware", "Software" ], "authors": [ { "affiliation": null, "fullName": "Blaise Tine", "givenName": "Blaise", "surname": "Tine", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Varun Saxena", "givenName": "Varun", "surname": "Saxena", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Santosh Srivatsan", "givenName": "Santosh", "surname": "Srivatsan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Joshua R. Simpson", "givenName": "Joshua R.", "surname": "Simpson", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Fadi Alzammar", "givenName": "Fadi", "surname": "Alzammar", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Liam Paul Cooper", "givenName": "Liam Paul", "surname": "Cooper", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Sam Jijina", "givenName": "Sam", "surname": "Jijina", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Swetha Rajagoplan", "givenName": "Swetha", "surname": "Rajagoplan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Tejaswini Anand Kumar", "givenName": "Tejaswini Anand", "surname": "Kumar", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jeff Young", "givenName": "Jeff", "surname": "Young", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Hyesoon Kim", "givenName": "Hyesoon", "surname": "Kim", "__typename": "ArticleAuthorType" } ], "idPrefix": "hcs", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-08-01T00:00:00", "pubType": "proceedings", "pages": "1-15", "year": "2022", "issn": null, "isbn": "978-1-6654-6028-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09895600", "articleId": "1GZiGtt4pgI", "__typename": "AdjacentArticleType" }, "next": { "fno": "09895613", "articleId": "1GZiFLRn9kc", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icie/2010/4080/2/05571266", "title": "Ambient Illumination Consistency Based on Hue Method in a Mixed-scene", "doi": null, "abstractUrl": "/proceedings-article/icie/2010/05571266/12OmNBQkwZz", "parentPublication": { "id": "proceedings/icie/2010/4080/2", "title": "Information Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2017/2219/0/2219a315", "title": "Accessing CUDA Features in the OpenGL Rendering Pipeline: A Case Study Using N-Body Simulation", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2017/2219a315/12OmNClQ0yg", "parentPublication": { "id": "proceedings/sibgrapi/2017/2219/0", "title": "2017 30th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pdcat/2008/3443/0/3443a143", "title": "Overheads in Accelerating Molecular Dynamics Simulations with GPUs", "doi": null, "abstractUrl": "/proceedings-article/pdcat/2008/3443a143/12OmNx57HQG", "parentPublication": { "id": "proceedings/pdcat/2008/3443/0", "title": "2008 Ninth International Conference on Parallel and Distributed Computing, Applications and Technologies", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cmpcon/1992/2655/0/00186697", "title": "Scalable graphics enhancements for PA-RISC workstations", "doi": null, "abstractUrl": "/proceedings-article/cmpcon/1992/00186697/12OmNxGSm2u", "parentPublication": { "id": "proceedings/cmpcon/1992/2655/0", "title": "COMPCON Spring 1992", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/comgeo/2014/4321/0/06910115", "title": "An Efficient Architecture for Automatic Shaders Management on Virtual Globes", "doi": null, "abstractUrl": "/proceedings-article/comgeo/2014/06910115/12OmNzd7bit", "parentPublication": { "id": "proceedings/comgeo/2014/4321/0", "title": "2014 5th International Conference on Computing for Geospatial Research and Application (COM.Geo)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/10/07378994", "title": "Compressed Coverage Masks for Path Rendering on Mobile GPUs", "doi": null, "abstractUrl": "/journal/tg/2016/10/07378994/13rRUwIF6l9", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/td/2019/02/08440753", "title": "Visibility Rendering Order: Improving Energy Efficiency on Mobile GPUs through Frame Coherence", "doi": null, "abstractUrl": "/journal/td/2019/02/08440753/17D45W1Oa5j", "parentPublication": { "id": "trans/td", "title": "IEEE Transactions on Parallel & Distributed Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hpca/2019/1444/0/144400a623", "title": "Rendering Elimination: Early Discard of Redundant Tiles in the Graphics Pipeline", "doi": null, "abstractUrl": "/proceedings-article/hpca/2019/144400a623/18M7PSwaQkE", "parentPublication": { "id": "proceedings/hpca/2019/1444/0", "title": "2019 IEEE International Symposium on High Performance Computer Architecture (HPCA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ipdps/2022/8106/0/810600a179", "title": "Top-Down Performance Profiling on NVIDIA&#x0027;s GPUs", "doi": null, "abstractUrl": "/proceedings-article/ipdps/2022/810600a179/1F1WcB7OVbO", "parentPublication": { "id": "proceedings/ipdps/2022/8106/0", "title": "2022 IEEE International Parallel and Distributed Processing Symposium (IPDPS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hpca/2021/2235/0/223500a709", "title": "CHOPIN: Scalable Graphics Rendering in Multi-GPU Systems via Parallel Image Composition", "doi": null, "abstractUrl": "/proceedings-article/hpca/2021/223500a709/1t0HYm80lLa", "parentPublication": { "id": "proceedings/hpca/2021/2235/0", "title": "2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1I0bRZfL0L6", "title": "SC22: International Conference for High Performance Computing, Networking, Storage and Analysis", "acronym": "sc", "groupId": "10046045", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1I0bSTFFUE8", "doi": "10.1109/SC41404.2022.00032", "title": "Accelerating Elliptic Curve Digital Signature Algorithms on GPUs", "normalizedTitle": "Accelerating Elliptic Curve Digital Signature Algorithms on GPUs", "abstract": "The Elliptic Curve Digital Signature Algorithm (ECDSA) is an essential building block of various cryptographic protocols. In particular, most blockchain systems adopt it to ensure transaction integrity. However, due to its high computational intensity, ECDSA is often the performance bottleneck in blockchain transaction processing. Recent work has accelerated ECDSA algorithms on the CPU; in contrast, success has been limited on the GPU, which has great potential for parallelization but is challenging for implementing elliptic curve functions. In this paper, we propose RapidEC, a GPU-based ECDSA implementation for SM2, a popular elliptic curve. Specifically, we design architecture-aware parallel primitives for elliptic curve point operations, and parallelize the processing of a single SM2 request as well as batches of requests. Consequently, our GPU-based RapidEC outperformed the state-of-the-art CPU-based algorithm by orders of magnitude. Additionally, our GPU-based modular arithmetic functions as well as point operation primitives can be applied to other computation tasks.", "abstracts": [ { "abstractType": "Regular", "content": "The Elliptic Curve Digital Signature Algorithm (ECDSA) is an essential building block of various cryptographic protocols. In particular, most blockchain systems adopt it to ensure transaction integrity. However, due to its high computational intensity, ECDSA is often the performance bottleneck in blockchain transaction processing. Recent work has accelerated ECDSA algorithms on the CPU; in contrast, success has been limited on the GPU, which has great potential for parallelization but is challenging for implementing elliptic curve functions. In this paper, we propose RapidEC, a GPU-based ECDSA implementation for SM2, a popular elliptic curve. Specifically, we design architecture-aware parallel primitives for elliptic curve point operations, and parallelize the processing of a single SM2 request as well as batches of requests. Consequently, our GPU-based RapidEC outperformed the state-of-the-art CPU-based algorithm by orders of magnitude. Additionally, our GPU-based modular arithmetic functions as well as point operation primitives can be applied to other computation tasks.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The Elliptic Curve Digital Signature Algorithm (ECDSA) is an essential building block of various cryptographic protocols. In particular, most blockchain systems adopt it to ensure transaction integrity. However, due to its high computational intensity, ECDSA is often the performance bottleneck in blockchain transaction processing. Recent work has accelerated ECDSA algorithms on the CPU; in contrast, success has been limited on the GPU, which has great potential for parallelization but is challenging for implementing elliptic curve functions. In this paper, we propose RapidEC, a GPU-based ECDSA implementation for SM2, a popular elliptic curve. Specifically, we design architecture-aware parallel primitives for elliptic curve point operations, and parallelize the processing of a single SM2 request as well as batches of requests. Consequently, our GPU-based RapidEC outperformed the state-of-the-art CPU-based algorithm by orders of magnitude. Additionally, our GPU-based modular arithmetic functions as well as point operation primitives can be applied to other computation tasks.", "fno": "544400a368", "keywords": [ "Blockchains", "Cryptographic Protocols", "Digital Signatures", "Graphics Processing Units", "Public Key Cryptography", "Transaction Processing", "Blockchain Transaction Processing", "Design Architecture Aware Parallel Primitives", "ECDSA Algorithms", "Elliptic Curve Digital Signature Algorithm", "Elliptic Curve Digital Signature Algorithms", "Elliptic Curve Functions", "Elliptic Curve Point Operations", "Essential Building Block", "GPU Based ECDSA Implementation", "GPU Based Modular Arithmetic Functions", "GPU Based Rapid EC", "Popular Elliptic Curve", "State Of The Art CPU Based Algorithm", "Elliptic Curves", "Source Coding", "Memory Management", "Graphics Processing Units", "Throughput", "Libraries", "Blockchains", "ECDSA", "Parallel Processing", "GPU" ], "authors": [ { "affiliation": "Hong Kong University of Science and Technology,Hong Kong SAR,China", "fullName": "Zonghao Feng", "givenName": "Zonghao", "surname": "Feng", "__typename": "ArticleAuthorType" }, { "affiliation": "Hong Kong University of Science and Technology,Hong Kong SAR,China", "fullName": "Qipeng Xie", "givenName": "Qipeng", "surname": "Xie", "__typename": "ArticleAuthorType" }, { "affiliation": "Hong Kong University of Science and Technology,Hong Kong SAR,China", "fullName": "Qiong Luo", "givenName": "Qiong", "surname": "Luo", "__typename": "ArticleAuthorType" }, { "affiliation": "WeBank,Shenzhen,China", "fullName": "Yujie Chen", "givenName": "Yujie", "surname": "Chen", "__typename": "ArticleAuthorType" }, { "affiliation": "WeBank,Shenzhen,China", "fullName": "Haoxuan Li", "givenName": "Haoxuan", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "WeBank,Shenzhen,China", "fullName": "Huizhong Li", "givenName": "Huizhong", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "WeBank,Shenzhen,China", "fullName": "Qiang Yan", "givenName": "Qiang", "surname": "Yan", "__typename": "ArticleAuthorType" } ], "idPrefix": "sc", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-11-01T00:00:00", "pubType": "proceedings", "pages": "1-13", "year": "2022", "issn": null, "isbn": "978-1-6654-5444-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "544400a480", "articleId": "1L07t3Tblde", "__typename": "AdjacentArticleType" }, "next": { "fno": "544400a495", "articleId": "1L07jJ9DvPO", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/euros&p/2018/4228/0/422801a384", "title": "In Search of CurveSwap: Measuring Elliptic Curve Implementations in the Wild", "doi": null, "abstractUrl": "/proceedings-article/euros&p/2018/422801a384/12OmNAgY7pI", "parentPublication": { "id": "proceedings/euros&p/2018/4228/0", "title": "2018 IEEE European Symposium on Security and Privacy (EuroS&P)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cis/2017/4822/0/482201a248", "title": "Practical Partial-Nonce-Exposure Attack on ECC Algorithm", "doi": null, "abstractUrl": "/proceedings-article/cis/2017/482201a248/12OmNBuL1jv", "parentPublication": { "id": "proceedings/cis/2017/4822/0", "title": "2017 13th International Conference on Computational Intelligence and Security (CIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sbac-pad/2017/1233/0/1233a033", "title": "Scalability of CPU and GPU Solutions of the Prime Elliptic Curve Discrete Logarithm Problem", "doi": null, "abstractUrl": "/proceedings-article/sbac-pad/2017/1233a033/12OmNqIzgSe", "parentPublication": { "id": "proceedings/sbac-pad/2017/1233/0", "title": "2017 29th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/eecs/2017/2085/0/2085a017", "title": "A New Algorithm for a Public Key Cryptosystem Using Elliptic Curve", "doi": null, "abstractUrl": "/proceedings-article/eecs/2017/2085a017/12OmNxZ2Gls", "parentPublication": { "id": "proceedings/eecs/2017/2085/0", "title": "2017 European Conference on Electrical Engineering and Computer Science (EECS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/lcn/2011/926/0/06115561", "title": "A light weight security scheme for HWMP protocol using Elliptic Curve technique", "doi": null, "abstractUrl": "/proceedings-article/lcn/2011/06115561/12OmNxwWoGk", "parentPublication": { "id": "proceedings/lcn/2011/926/0", "title": "2011 IEEE 36th Conference on Local Computer Networks", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/asap/2010/6966/0/05541000", "title": "Elliptic Curve point multiplication on GPUs", "doi": null, "abstractUrl": "/proceedings-article/asap/2010/05541000/12OmNyv7m3v", "parentPublication": { "id": "proceedings/asap/2010/6966/0", "title": "ASAP 2010 - 21st IEEE International Conference on Application-specific Systems, Architectures and Processors", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sc/2022/5444/0/544400a368", "title": "Accelerating Elliptic Curve Digital Signature Algorithms on GPUs", "doi": null, "abstractUrl": "/proceedings-article/sc/2022/544400a368/1L07usvop3O", "parentPublication": { "id": "proceedings/sc/2022/5444/0/", "title": "SC22: International Conference for High Performance Computing, Networking, Storage and Analysis", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/euros&p/2019/1148/0/114800a371", "title": "Degenerate Fault Attacks on Elliptic Curve Parameters in OpenSSL", "doi": null, "abstractUrl": "/proceedings-article/euros&p/2019/114800a371/1cI6fJL13Vu", "parentPublication": { "id": "proceedings/euros&p/2019/1148/0", "title": "2019 IEEE European Symposium on Security and Privacy (EuroS&P)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cis/2019/6092/0/609200a252", "title": "Secret Sharing Based SM2 Digital Signature Generation using Homomorphic Encryption", "doi": null, "abstractUrl": "/proceedings-article/cis/2019/609200a252/1i5m4iTtom4", "parentPublication": { "id": "proceedings/cis/2019/6092/0", "title": "2019 15th International Conference on Computational Intelligence and Security (CIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nana/2021/4158/0/415800a443", "title": "Signature Scheme Based on The SM2 Algorithm in Fabric", "doi": null, "abstractUrl": "/proceedings-article/nana/2021/415800a443/1zdPJBHmwus", "parentPublication": { "id": "proceedings/nana/2021/4158/0", "title": "2021 International Conference on Networking and Network Applications (NaNA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1qRNrlo577W", "title": "2020 IEEE Visualization Conference (VIS)", "acronym": "vis", "groupId": "1001944", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1qROP1sMu2I", "doi": "10.1109/VIS47514.2020.00009", "title": "A Virtual Frame Buffer Abstraction for Parallel Rendering of Large Tiled Display Walls", "normalizedTitle": "A Virtual Frame Buffer Abstraction for Parallel Rendering of Large Tiled Display Walls", "abstract": "We present dw2, a flexible and easy-to-use software infrastructure for interactive rendering of large tiled display walls. Our library represents the tiled display wall as a single virtual screen through a display \"service\", which renderers connect to and send image tiles to be displayed, either from an on-site or remote cluster. The display service can be easily configured to support a range of typical network and display hardware configurations; the client library provides a straightforward interface for easy integration into existing renderers. We evaluate the performance of our display wall service in different configurations using a CPU and GPU ray tracer, in both on-site and remote rendering scenarios using multiple display walls.", "abstracts": [ { "abstractType": "Regular", "content": "We present dw2, a flexible and easy-to-use software infrastructure for interactive rendering of large tiled display walls. Our library represents the tiled display wall as a single virtual screen through a display \"service\", which renderers connect to and send image tiles to be displayed, either from an on-site or remote cluster. The display service can be easily configured to support a range of typical network and display hardware configurations; the client library provides a straightforward interface for easy integration into existing renderers. We evaluate the performance of our display wall service in different configurations using a CPU and GPU ray tracer, in both on-site and remote rendering scenarios using multiple display walls.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We present dw2, a flexible and easy-to-use software infrastructure for interactive rendering of large tiled display walls. Our library represents the tiled display wall as a single virtual screen through a display \"service\", which renderers connect to and send image tiles to be displayed, either from an on-site or remote cluster. The display service can be easily configured to support a range of typical network and display hardware configurations; the client library provides a straightforward interface for easy integration into existing renderers. We evaluate the performance of our display wall service in different configurations using a CPU and GPU ray tracer, in both on-site and remote rendering scenarios using multiple display walls.", "fno": "801400a011", "keywords": [ "Buffer Storage", "Computer Displays", "Data Visualisation", "Graphics Processing Units", "Interactive Devices", "Microprocessor Chips", "Parallel Processing", "Pattern Clustering", "Ray Tracing", "Rendering Computer Graphics", "Virtual Frame Buffer Abstraction", "Parallel Rendering", "Tiled Display Wall", "Software Infrastructure", "Interactive Rendering", "Image Tiles", "Display Wall Service", "Multiple Display Walls", "GPU Ray Tracer", "CPU Ray Tracer", "Visualization", "Conferences", "Graphics Processing Units", "Rendering Computer Graphics", "Libraries", "Software", "Hardware", "Tiled Display Walls", "Distributed Display Frameworks" ], "authors": [ { "affiliation": "University of Utah,SCI Institute", "fullName": "Mengjiao Han", "givenName": "Mengjiao", "surname": "Han", "__typename": "ArticleAuthorType" }, { "affiliation": "NVIDIA Corp", "fullName": "Ingo Wald", "givenName": "Ingo", "surname": "Wald", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Utah,SCI Institute", "fullName": "Will Usher", "givenName": "Will", "surname": "Usher", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Utah,SCI Institute", "fullName": "Nate Morrical", "givenName": "Nate", "surname": "Morrical", "__typename": "ArticleAuthorType" }, { "affiliation": "Intel Corp.", "fullName": "Aaron Knoll", "givenName": "Aaron", "surname": "Knoll", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Utah,SCI Institute", "fullName": "Valerio Pascucci", "givenName": "Valerio", "surname": "Pascucci", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Utah,SCI Institute", "fullName": "Chris R. Johnson", "givenName": "Chris R.", "surname": "Johnson", "__typename": "ArticleAuthorType" } ], "idPrefix": "vis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-10-01T00:00:00", "pubType": "proceedings", "pages": "11-15", "year": "2020", "issn": null, "isbn": "978-1-7281-8014-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1qROMnkEUVO", "name": "pvis202080140-09331308s1-mm_801400a011.zip", "size": "92.9 MB", "location": "https://www.computer.org/csdl/api/v1/extra/pvis202080140-09331308s1-mm_801400a011.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "801400a006", "articleId": "1qRNs1dVBJu", "__typename": "AdjacentArticleType" }, "next": { "fno": "801400a016", "articleId": "1qRORkoFgUU", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/e-science/2013/5083/0/06683916", "title": "Magic View: An Optimized Ultra-Large Scientific Image Viewer for SAGE Tiled-Display Environment", "doi": null, "abstractUrl": "/proceedings-article/e-science/2013/06683916/12OmNASraXD", "parentPublication": { "id": "proceedings/e-science/2013/5083/0", "title": "2013 IEEE 9th International Conference on eScience (eScience)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icig/2004/2244/0/01410480", "title": "A survey of multi-projector tiled display wall construction", "doi": null, "abstractUrl": "/proceedings-article/icig/2004/01410480/12OmNAWH9up", "parentPublication": { "id": "proceedings/icig/2004/2244/0", "title": "Proceedings. Third International Conference on Image and Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nbis/2013/2510/0/2510a305", "title": "Proposal of Disaster Management Support System Using Tiled Display Wall Environment", "doi": null, "abstractUrl": "/proceedings-article/nbis/2013/2510a305/12OmNApcuB9", "parentPublication": { "id": "proceedings/nbis/2013/2510/0", "title": "2013 16th International Conference on Network-Based Information Systems (NBiS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nbis/2011/4458/0/4458a456", "title": "A New Interface for Large Scale Tiled Display System Considering Scalability", "doi": null, "abstractUrl": "/proceedings-article/nbis/2011/4458a456/12OmNBU1jQj", "parentPublication": { "id": "proceedings/nbis/2011/4458/0", "title": "2011 14th International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nbis/2012/4779/0/4779a540", "title": "Approaches to Display of Ultra-resolution Video Streaming by Multi-transmission on Tiled Display Environment", "doi": null, "abstractUrl": "/proceedings-article/nbis/2012/4779a540/12OmNqOOrKv", "parentPublication": { "id": "proceedings/nbis/2012/4779/0", "title": "2012 15th International Conference on Network-Based Information Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ldav/2014/5215/0/07013217", "title": "Remote parallel rendering for high-resolution tiled display walls", "doi": null, "abstractUrl": "/proceedings-article/ldav/2014/07013217/12OmNwdtw8h", "parentPublication": { "id": "proceedings/ldav/2014/5215/0", "title": "2014 IEEE 4th Symposium on Large Data Analysis and Visualization (LDAV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icppw/2011/4511/0/4511a308", "title": "Integrating Scientific Workflows and Large Tiled Display Walls: Bridging the Visualization Divide", "doi": null, "abstractUrl": "/proceedings-article/icppw/2011/4511a308/12OmNxE2mMD", "parentPublication": { "id": "proceedings/icppw/2011/4511/0", "title": "2011 40th International Conference on Parallel Processing Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/05/v0864", "title": "Garuda: A Scalable Tiled Display Wall Using Commodity PCs", "doi": null, "abstractUrl": "/journal/tg/2007/05/v0864/13rRUwgQpqF", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2005/04/mcg2005040024", "title": "Tools and Applications for Large-Scale Display Walls", "doi": null, "abstractUrl": "/magazine/cg/2005/04/mcg2005040024/13rRUxYINaF", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2010/06/ttg2010061623", "title": "A Scalable Distributed Paradigm for Multi-User Interaction with Tiled Rear Projection Display Walls", "doi": null, "abstractUrl": "/journal/tg/2010/06/ttg2010061623/13rRUxYrbUB", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }