data
dict
{ "proceeding": { "id": "12OmNro0Iar", "title": "Computer Science and Electronics Engineering, International Conference on", "acronym": "iccsee", "groupId": "1801147", "volume": "3", "displayVolume": "3", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNAXxXhU", "doi": "10.1109/ICCSEE.2012.129", "title": "A Survey of Computer Facial Animation Techniques", "normalizedTitle": "A Survey of Computer Facial Animation Techniques", "abstract": "This paper describes and surveys the techniques used in facial animation. The techniques are discussed from two aspects: the techniques of facial modeling and the techniques of data acquirement of facial animation. The techniques of facial modeling are classified into four categories: shape interpolations, parameterizations, muscle-based modeling and pseudo-muscle-based models. And the techniques of animation data acquirement are classified into three categories: speech-driven techniques, image-based techniques and data-capture techniques. The generation, the main ideas, the historical use, and the strength and weakness of each technique are described in detail.", "abstracts": [ { "abstractType": "Regular", "content": "This paper describes and surveys the techniques used in facial animation. The techniques are discussed from two aspects: the techniques of facial modeling and the techniques of data acquirement of facial animation. The techniques of facial modeling are classified into four categories: shape interpolations, parameterizations, muscle-based modeling and pseudo-muscle-based models. And the techniques of animation data acquirement are classified into three categories: speech-driven techniques, image-based techniques and data-capture techniques. The generation, the main ideas, the historical use, and the strength and weakness of each technique are described in detail.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper describes and surveys the techniques used in facial animation. The techniques are discussed from two aspects: the techniques of facial modeling and the techniques of data acquirement of facial animation. The techniques of facial modeling are classified into four categories: shape interpolations, parameterizations, muscle-based modeling and pseudo-muscle-based models. And the techniques of animation data acquirement are classified into three categories: speech-driven techniques, image-based techniques and data-capture techniques. The generation, the main ideas, the historical use, and the strength and weakness of each technique are described in detail.", "fno": "4647c434", "keywords": [ "Facial Animation", "Facial Modeling", "Animation Data Acquirement" ], "authors": [ { "affiliation": null, "fullName": "Linghua Li", "givenName": "Linghua", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yongkui Liu", "givenName": "Yongkui", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Hengbo Zhang", "givenName": "Hengbo", "surname": "Zhang", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccsee", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-03-01T00:00:00", "pubType": "proceedings", "pages": "434-438", "year": "2012", "issn": null, "isbn": "978-0-7695-4647-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4647c430", "articleId": "12OmNAolGZs", "__typename": "AdjacentArticleType" }, "next": { "fno": "4647c439", "articleId": "12OmNxymo6y", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/smi/2002/1546/0/15460219", "title": "Statistical Generation of 3D Facial Animable Models", "doi": null, "abstractUrl": "/proceedings-article/smi/2002/15460219/12OmNAle6uJ", "parentPublication": { "id": "proceedings/smi/2002/1546/0", "title": "Proceedings SMI. Shape Modeling International 2002", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/kse/2009/3846/0/3846a081", "title": "Fast and Realistic 2D Facial Animation Based on Image Warping", "doi": null, "abstractUrl": "/proceedings-article/kse/2009/3846a081/12OmNqGA59e", "parentPublication": { "id": "proceedings/kse/2009/3846/0", "title": "Knowledge and Systems Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/1996/7588/0/75880098", "title": "Facial Animation", "doi": null, "abstractUrl": "/proceedings-article/ca/1996/75880098/12OmNvT2oR2", "parentPublication": { "id": "proceedings/ca/1996/7588/0", "title": "Computer Animation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/1996/7588/0/75880068", "title": "Modeling, Tracking and Interactive Animation of Faces and Heads Using Input from Video", "doi": null, "abstractUrl": "/proceedings-article/ca/1996/75880068/12OmNwfKjaJ", "parentPublication": { "id": "proceedings/ca/1996/7588/0", "title": "Computer Animation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/2001/7237/0/00982374", "title": "A physically-based model with adaptive refinement for facial animation", "doi": null, "abstractUrl": "/proceedings-article/ca/2001/00982374/12OmNxRWI7R", "parentPublication": { "id": "proceedings/ca/2001/7237/0", "title": "Proceedings Computer Animation 2001. Fourteenth Conference on Computer Animation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2014/4761/0/06890231", "title": "Real-time control of 3D facial animation", "doi": null, "abstractUrl": "/proceedings-article/icme/2014/06890231/12OmNyOHG1A", "parentPublication": { "id": "proceedings/icme/2014/4761/0", "title": "2014 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2010/4215/0/4215a425", "title": "Computer Animation of Facial Emotions", "doi": null, "abstractUrl": "/proceedings-article/cw/2010/4215a425/12OmNzTYC9m", "parentPublication": { "id": "proceedings/cw/2010/4215/0", "title": "2010 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2010/04/mcg2010040051", "title": "Modeling Short-Term Dynamics and Variability for Realistic Interactive Facial Animation", "doi": null, "abstractUrl": "/magazine/cg/2010/04/mcg2010040051/13rRUwgQpwW", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2004/03/v0339", "title": "A New Physical Model with Multilayer Architecture for Facial Expression Animation Using Dynamic Adaptive Mesh", "doi": null, "abstractUrl": "/journal/tg/2004/03/v0339/13rRUxD9gXw", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2005/03/v0341", "title": "Creating Speech-Synchronized Animation", "doi": null, "abstractUrl": "/journal/tg/2005/03/v0341/13rRUxE04tq", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNB8Cj92", "title": "2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)", "acronym": "icmew", "groupId": "1801805", "volume": "0", "displayVolume": "0", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNBC8Ayh", "doi": "10.1109/ICMEW.2014.6890554", "title": "Realtime speech-driven facial animation using Gaussian Mixture Models", "normalizedTitle": "Realtime speech-driven facial animation using Gaussian Mixture Models", "abstract": "Synthesizing speech-driven facial animation is the process of animating a virtual face according to the input audio signal. Actually, audio-to-visual conversion is the core of speech-driven facial animation. In this paper, Gaussian Mixture Models (GMM) are employed for audio-to-visual conversion. The conventional GMM based method performs the conversion frame by frame using minimum mean square error estimation. We consider two issues related to the conventional method: 1) the influence of previous visual features on current visual feature is not considered, and 2) GMM training and conversion are inconsistent. To address these issues, we propose incorporating previous visual features into the conversion. We also propose a minimum conversion error based approach to refine the GMM parameters. Experiments on a public available database show that our method can accurately convert audio features into visual features. The conversion accuracy is comparable to a current state-of-the-art trajectory-based approach. Based on the proposed method, we develop a speech-driven facial animation system, the system runs in real time and outputs realistic speech animations.", "abstracts": [ { "abstractType": "Regular", "content": "Synthesizing speech-driven facial animation is the process of animating a virtual face according to the input audio signal. Actually, audio-to-visual conversion is the core of speech-driven facial animation. In this paper, Gaussian Mixture Models (GMM) are employed for audio-to-visual conversion. The conventional GMM based method performs the conversion frame by frame using minimum mean square error estimation. We consider two issues related to the conventional method: 1) the influence of previous visual features on current visual feature is not considered, and 2) GMM training and conversion are inconsistent. To address these issues, we propose incorporating previous visual features into the conversion. We also propose a minimum conversion error based approach to refine the GMM parameters. Experiments on a public available database show that our method can accurately convert audio features into visual features. The conversion accuracy is comparable to a current state-of-the-art trajectory-based approach. Based on the proposed method, we develop a speech-driven facial animation system, the system runs in real time and outputs realistic speech animations.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Synthesizing speech-driven facial animation is the process of animating a virtual face according to the input audio signal. Actually, audio-to-visual conversion is the core of speech-driven facial animation. In this paper, Gaussian Mixture Models (GMM) are employed for audio-to-visual conversion. The conventional GMM based method performs the conversion frame by frame using minimum mean square error estimation. We consider two issues related to the conventional method: 1) the influence of previous visual features on current visual feature is not considered, and 2) GMM training and conversion are inconsistent. To address these issues, we propose incorporating previous visual features into the conversion. We also propose a minimum conversion error based approach to refine the GMM parameters. Experiments on a public available database show that our method can accurately convert audio features into visual features. The conversion accuracy is comparable to a current state-of-the-art trajectory-based approach. Based on the proposed method, we develop a speech-driven facial animation system, the system runs in real time and outputs realistic speech animations.", "fno": "06890554", "keywords": [ "Visualization", "Vectors", "Training", "Shape", "Facial Animation", "Principal Component Analysis", "GMM", "Facial Animation", "Speech Driven", "Audio To Visual Conversion" ], "authors": [ { "affiliation": "Department of Automation, University of Science and Technology of China, China", "fullName": "Changwei Luo", "givenName": null, "surname": "Changwei Luo", "__typename": "ArticleAuthorType" }, { "affiliation": "Department of Automation, University of Science and Technology of China, China", "fullName": "Jun Yu", "givenName": null, "surname": "Jun Yu", "__typename": "ArticleAuthorType" }, { "affiliation": "Department of Automation, University of Science and Technology of China, China", "fullName": "Xian Li", "givenName": null, "surname": "Xian Li", "__typename": "ArticleAuthorType" }, { "affiliation": "Department of Automation, University of Science and Technology of China, China", "fullName": "Zengfu Wang", "givenName": null, "surname": "Zengfu Wang", "__typename": "ArticleAuthorType" } ], "idPrefix": "icmew", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-07-01T00:00:00", "pubType": "proceedings", "pages": "1-6", "year": "2014", "issn": "1945-7871", "isbn": "978-1-4799-4717-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06890553", "articleId": "12OmNrkBwlf", "__typename": "AdjacentArticleType" }, "next": { "fno": "06890555", "articleId": "12OmNwHQB9E", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccsee/2012/4647/3/4647c434", "title": "A Survey of Computer Facial Animation Techniques", "doi": null, "abstractUrl": "/proceedings-article/iccsee/2012/4647c434/12OmNAXxXhU", "parentPublication": { "id": "proceedings/iccsee/2012/4647/3", "title": "Computer Science and Electronics Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pg/2002/1784/0/17840077", "title": "\"May I talk to you? :-)\" — Facial Animation from Text", "doi": null, "abstractUrl": "/proceedings-article/pg/2002/17840077/12OmNAkWveH", "parentPublication": { "id": "proceedings/pg/2002/1784/0", "title": "Computer Graphics and Applications, Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icassp/2001/7041/2/00941046", "title": "Voice conversion algorithm based on Gaussian mixture model with dynamic frequency warping of STRAIGHT spectrum", "doi": null, "abstractUrl": "/proceedings-article/icassp/2001/00941046/12OmNwNeYBW", "parentPublication": { "id": "proceedings/icassp/2001/7041/2", "title": "Acoustics, Speech, and Signal Processing, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2017/0733/0/0733c328", "title": "Speech-Driven 3D Facial Animation with Implicit Emotional Awareness: A Deep Learning Approach", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2017/0733c328/12OmNxE2mG1", "parentPublication": { "id": "proceedings/cvprw/2017/0733/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2006/2521/1/252111128", "title": "Speech Animation Using Coupled Hidden Markov Models", "doi": null, "abstractUrl": "/proceedings-article/icpr/2006/252111128/12OmNy2agWk", "parentPublication": { "id": "proceedings/icpr/2006/2521/1", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icassp/2004/8484/3/01326624", "title": "Inner lip feature extraction for MPEG-4 facial animation", "doi": null, "abstractUrl": "/proceedings-article/icassp/2004/01326624/12OmNyen1ka", "parentPublication": { "id": "proceedings/icassp/2004/8484/3", "title": "2004 IEEE International Conference on Acoustics, Speech, and Signal Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/elmar/2006/4403/0/04127510", "title": "Database Construction for Speech to Lip-readable Animation Conversion", "doi": null, "abstractUrl": "/proceedings-article/elmar/2006/04127510/12OmNzuZUAs", "parentPublication": { "id": "proceedings/elmar/2006/4403/0", "title": "International Symposium ELMAR", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2006/06/v1523", "title": "Expressive Facial Animation Synthesis by Learning Speech Coarticulation and Expression Spaces", "doi": null, "abstractUrl": "/journal/tg/2006/06/v1523/13rRUxASubv", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09992151", "title": "Personalized Audio-Driven 3D Facial Animation Via Style-Content Disentanglement", "doi": null, "abstractUrl": "/journal/tg/5555/01/09992151/1JevBLSiUqA", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ta/2022/03/09140332", "title": "Emotion Dependent Domain Adaptation for Speech Driven Affective Facial Feature Synthesis", "doi": null, "abstractUrl": "/journal/ta/2022/03/09140332/1lsnzQkrydG", "parentPublication": { "id": "trans/ta", "title": "IEEE Transactions on Affective Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxETa76", "title": "Computer Animation", "acronym": "ca", "groupId": "1000121", "volume": "0", "displayVolume": "0", "year": "1996", "__typename": "ProceedingType" }, "article": { "id": "12OmNvT2oR2", "doi": "10.1109/CA.1996.540492", "title": "Facial Animation", "normalizedTitle": "Facial Animation", "abstract": "This paper present an application to the facial animation of the D.O.G.M.A. (Deformation Of Geometrical Model Animated) model. We described an order-controlled animation, where a first set of orders simulate the contraction of different facial muscles, a second some facial expressions. The main interest of this method is the manipulation of faces with various shape. Since the deformation given by the animation model DOGMA are space deformations, the only necessity is a hierarchical structuring of the different face component. With this method the same facial animation order can easily be used for various objects with different shapes, if they present the same hierarchical structuring.", "abstracts": [ { "abstractType": "Regular", "content": "This paper present an application to the facial animation of the D.O.G.M.A. (Deformation Of Geometrical Model Animated) model. We described an order-controlled animation, where a first set of orders simulate the contraction of different facial muscles, a second some facial expressions. The main interest of this method is the manipulation of faces with various shape. Since the deformation given by the animation model DOGMA are space deformations, the only necessity is a hierarchical structuring of the different face component. With this method the same facial animation order can easily be used for various objects with different shapes, if they present the same hierarchical structuring.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper present an application to the facial animation of the D.O.G.M.A. (Deformation Of Geometrical Model Animated) model. We described an order-controlled animation, where a first set of orders simulate the contraction of different facial muscles, a second some facial expressions. The main interest of this method is the manipulation of faces with various shape. Since the deformation given by the animation model DOGMA are space deformations, the only necessity is a hierarchical structuring of the different face component. With this method the same facial animation order can easily be used for various objects with different shapes, if they present the same hierarchical structuring.", "fno": "75880098", "keywords": [ "Animation", "Deformation", "Facial Animation", "Order Controlled" ], "authors": [ { "affiliation": "LSIIT", "fullName": "Nicolas Dubreuil", "givenName": "Nicolas", "surname": "Dubreuil", "__typename": "ArticleAuthorType" }, { "affiliation": "LSIIT", "fullName": "Dominique Bechmann", "givenName": "Dominique", "surname": "Bechmann", "__typename": "ArticleAuthorType" } ], "idPrefix": "ca", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1996-06-01T00:00:00", "pubType": "proceedings", "pages": "98", "year": "1996", "issn": "1087-4844", "isbn": "0-8186-7588-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "75880090", "articleId": "12OmNzYeAO0", "__typename": "AdjacentArticleType" }, "next": { "fno": "75880110", "articleId": "12OmNyNQSPN", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAkEU4f", "title": "2011 IEEE International Conference on Multimedia and Expo", "acronym": "icme", "groupId": "1000477", "volume": "0", "displayVolume": "0", "year": "2011", "__typename": "ProceedingType" }, "article": { "id": "12OmNviZlAw", "doi": "10.1109/ICME.2011.6011861", "title": "Animation of generic 3D head models driven by speech", "normalizedTitle": "Animation of generic 3D head models driven by speech", "abstract": "In this paper, a system for speech-driven animation of generic 3D head models is presented. The system is based on the inversion of a joint Audio-Visual Hidden Markov Model to estimate the visual information from speech data. Estimated visual speech features are used to animate a simple face model. The animation of a more complex head model is then obtained by automatically mapping the deformation of the simple model to it. The proposed algorithm allows the animation of 3D head models of arbitrary complexity through a simple setup procedure. The resulting animation is evaluated in terms of intelligibility of visual speech through subjective tests, showing a promising performance.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, a system for speech-driven animation of generic 3D head models is presented. The system is based on the inversion of a joint Audio-Visual Hidden Markov Model to estimate the visual information from speech data. Estimated visual speech features are used to animate a simple face model. The animation of a more complex head model is then obtained by automatically mapping the deformation of the simple model to it. The proposed algorithm allows the animation of 3D head models of arbitrary complexity through a simple setup procedure. The resulting animation is evaluated in terms of intelligibility of visual speech through subjective tests, showing a promising performance.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, a system for speech-driven animation of generic 3D head models is presented. The system is based on the inversion of a joint Audio-Visual Hidden Markov Model to estimate the visual information from speech data. Estimated visual speech features are used to animate a simple face model. The animation of a more complex head model is then obtained by automatically mapping the deformation of the simple model to it. The proposed algorithm allows the animation of 3D head models of arbitrary complexity through a simple setup procedure. The resulting animation is evaluated in terms of intelligibility of visual speech through subjective tests, showing a promising performance.", "fno": "06011861", "keywords": [ "Hidden Markov Models", "Visualization", "Speech", "Animation", "Face", "Feature Extraction", "Adaptation Models", "Facial Animation", "Hidden Markov Models", "Audio Visual Speech Processing" ], "authors": [ { "affiliation": "Lab. for System Dyn. & Signal Processing, Universidad Nacional de Rosario, CIFASIS, Argentina", "fullName": "Lucas Terissi", "givenName": "Lucas", "surname": "Terissi", "__typename": "ArticleAuthorType" }, { "affiliation": "Loria - INRIA Nancy Grand Est, Cortex Team, Vandoeuvre-lès, France", "fullName": "Mauricio Cerda", "givenName": "Mauricio", "surname": "Cerda", "__typename": "ArticleAuthorType" }, { "affiliation": "Lab. for System Dyn. & Signal Processing, Universidad Nacional de Rosario, CIFASIS, Argentina", "fullName": "Juan C. Gómez", "givenName": "Juan C.", "surname": "Gómez", "__typename": "ArticleAuthorType" }, { "affiliation": "Computer Science Department, FCFyM, Universidad de Chile, Santiago, Chile", "fullName": "Nancy Hitschfeld-Kahler", "givenName": "Nancy", "surname": "Hitschfeld-Kahler", "__typename": "ArticleAuthorType" }, { "affiliation": "Loria - INRIA Nancy Grand Est, Cortex Team, Vandoeuvre-lès, France", "fullName": "Bernard Girau", "givenName": "Bernard", "surname": "Girau", "__typename": "ArticleAuthorType" }, { "affiliation": "Computer Science Department, FCFyM, Universidad de Chile, Santiago, Chile", "fullName": "Renato Valenzuela", "givenName": "Renato", "surname": "Valenzuela", "__typename": "ArticleAuthorType" } ], "idPrefix": "icme", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2011-07-01T00:00:00", "pubType": "proceedings", "pages": "1-6", "year": "2011", "issn": "1945-7871", "isbn": "978-1-61284-348-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06011860", "articleId": "12OmNzZEADg", "__typename": "AdjacentArticleType" }, "next": { "fno": "06011899", "articleId": "12OmNrIJqqV", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/culture-computing/2011/4546/0/4546a121", "title": "Thai Speech-Driven Facial Animation", "doi": null, "abstractUrl": "/proceedings-article/culture-computing/2011/4546a121/12OmNARiM0A", "parentPublication": { "id": "proceedings/culture-computing/2011/4546/0", "title": "International Conference on Culture and Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2017/6067/0/08019546", "title": "Visual speech synthesis from 3D mesh sequences driven by combined speech features", "doi": null, "abstractUrl": "/proceedings-article/icme/2017/08019546/12OmNqBtiXN", "parentPublication": { "id": "proceedings/icme/2017/6067/0", "title": "2017 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iih-msp/2008/3278/0/3278a113", "title": "An Approach to Speech Driven Animation", "doi": null, "abstractUrl": "/proceedings-article/iih-msp/2008/3278a113/12OmNvw2TcQ", "parentPublication": { "id": "proceedings/iih-msp/2008/3278/0", "title": "2008 Fourth International Conference on Intelligent Information Hiding and Multimedia Signal Processing (IIH-MSP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2017/0733/0/0733c328", "title": "Speech-Driven 3D Facial Animation with Implicit Emotional Awareness: A Deep Learning Approach", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2017/0733c328/12OmNxE2mG1", "parentPublication": { "id": "proceedings/cvprw/2017/0733/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvmp/2010/4268/0/4268a171", "title": "Prominence Driven Character Animation", "doi": null, "abstractUrl": "/proceedings-article/cvmp/2010/4268a171/12OmNxFaLtB", "parentPublication": { "id": "proceedings/cvmp/2010/4268/0", "title": "2010 Conference on Visual Media Production", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/elmar/2006/4403/0/04127510", "title": "Database Construction for Speech to Lip-readable Animation Conversion", "doi": null, "abstractUrl": "/proceedings-article/elmar/2006/04127510/12OmNzuZUAs", "parentPublication": { "id": "proceedings/elmar/2006/4403/0", "title": "International Symposium ELMAR", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/11/ttg2012111915", "title": "A Statistical Quality Model for Data-Driven Speech Animation", "doi": null, "abstractUrl": "/journal/tg/2012/11/ttg2012111915/13rRUIIVlkf", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2011/05/mcg2011050080", "title": "Carnival—Combining Speech Technology and Computer Animation", "doi": null, "abstractUrl": "/magazine/cg/2011/05/mcg2011050080/13rRUxAASMX", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2006/06/v1523", "title": "Expressive Facial Animation Synthesis by Learning Speech Coarticulation and Expression Spaces", "doi": null, "abstractUrl": "/journal/tg/2006/06/v1523/13rRUxASubv", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ta/2022/03/09140332", "title": "Emotion Dependent Domain Adaptation for Speech Driven Affective Facial Feature Synthesis", "doi": null, "abstractUrl": "/journal/ta/2022/03/09140332/1lsnzQkrydG", "parentPublication": { "id": "trans/ta", "title": "IEEE Transactions on Affective Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAR1b0Z", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "acronym": "cvprw", "groupId": "1001809", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNxE2mG1", "doi": "10.1109/CVPRW.2017.287", "title": "Speech-Driven 3D Facial Animation with Implicit Emotional Awareness: A Deep Learning Approach", "normalizedTitle": "Speech-Driven 3D Facial Animation with Implicit Emotional Awareness: A Deep Learning Approach", "abstract": "We introduce a long short-term memory recurrent neural network (LSTM-RNN) approach for real-time facial animation, which automatically estimates head rotation and facial action unit activations of a speaker from just her speech. Specifically, the time-varying contextual non-linear mapping between audio stream and visual facial movements is realized by training a LSTM neural network on a large audio-visual data corpus. In this work, we extract a set of acoustic features from input audio, including Mel-scaled spectrogram, Mel frequency cepstral coefficients and chromagram that can effectively represent both contextual progression and emotional intensity of the speech. Output facial movements are characterized by 3D rotation and blending expression weights of a blendshape model, which can be used directly for animation. Thus, even though our model does not explicitly predict the affective states of the target speaker, her emotional manifestation is recreated via expression weights of the face model. Experiments on an evaluation dataset of different speakers across a wide range of affective states demonstrate promising results of our approach in real-time speech-driven facial animation.", "abstracts": [ { "abstractType": "Regular", "content": "We introduce a long short-term memory recurrent neural network (LSTM-RNN) approach for real-time facial animation, which automatically estimates head rotation and facial action unit activations of a speaker from just her speech. Specifically, the time-varying contextual non-linear mapping between audio stream and visual facial movements is realized by training a LSTM neural network on a large audio-visual data corpus. In this work, we extract a set of acoustic features from input audio, including Mel-scaled spectrogram, Mel frequency cepstral coefficients and chromagram that can effectively represent both contextual progression and emotional intensity of the speech. Output facial movements are characterized by 3D rotation and blending expression weights of a blendshape model, which can be used directly for animation. Thus, even though our model does not explicitly predict the affective states of the target speaker, her emotional manifestation is recreated via expression weights of the face model. Experiments on an evaluation dataset of different speakers across a wide range of affective states demonstrate promising results of our approach in real-time speech-driven facial animation.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We introduce a long short-term memory recurrent neural network (LSTM-RNN) approach for real-time facial animation, which automatically estimates head rotation and facial action unit activations of a speaker from just her speech. Specifically, the time-varying contextual non-linear mapping between audio stream and visual facial movements is realized by training a LSTM neural network on a large audio-visual data corpus. In this work, we extract a set of acoustic features from input audio, including Mel-scaled spectrogram, Mel frequency cepstral coefficients and chromagram that can effectively represent both contextual progression and emotional intensity of the speech. Output facial movements are characterized by 3D rotation and blending expression weights of a blendshape model, which can be used directly for animation. Thus, even though our model does not explicitly predict the affective states of the target speaker, her emotional manifestation is recreated via expression weights of the face model. Experiments on an evaluation dataset of different speakers across a wide range of affective states demonstrate promising results of our approach in real-time speech-driven facial animation.", "fno": "0733c328", "keywords": [ "Three Dimensional Displays", "Speech", "Feature Extraction", "Face", "Solid Modeling", "Hidden Markov Models", "Facial Animation" ], "authors": [ { "affiliation": null, "fullName": "Hai X. Pham", "givenName": "Hai X.", "surname": "Pham", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Samuel Cheung", "givenName": "Samuel", "surname": "Cheung", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Vladimir Pavlovic", "givenName": "Vladimir", "surname": "Pavlovic", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvprw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-07-01T00:00:00", "pubType": "proceedings", "pages": "2328-2336", "year": "2017", "issn": "2160-7516", "isbn": "978-1-5386-0733-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0733c318", "articleId": "12OmNyFU76C", "__typename": "AdjacentArticleType" }, "next": { "fno": "0733c337", "articleId": "12OmNzcPAaC", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccsee/2012/4647/3/4647c434", "title": "A Survey of Computer Facial Animation Techniques", "doi": null, "abstractUrl": "/proceedings-article/iccsee/2012/4647c434/12OmNAXxXhU", "parentPublication": { "id": "proceedings/iccsee/2012/4647/3", "title": "Computer Science and Electronics Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pg/2002/1784/0/17840077", "title": "\"May I talk to you? :-)\" — Facial Animation from Text", "doi": null, "abstractUrl": "/proceedings-article/pg/2002/17840077/12OmNAkWveH", "parentPublication": { "id": "proceedings/pg/2002/1784/0", "title": "Computer Graphics and Applications, Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2014/4717/0/06890554", "title": "Realtime speech-driven facial animation using Gaussian Mixture Models", "doi": null, "abstractUrl": "/proceedings-article/icmew/2014/06890554/12OmNBC8Ayh", "parentPublication": { "id": "proceedings/icmew/2014/4717/0", "title": "2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dmdcm/2011/4413/0/4413a132", "title": "Towards 3D Communications: Real Time Emotion Driven 3D Virtual Facial Animation", "doi": null, "abstractUrl": "/proceedings-article/dmdcm/2011/4413a132/12OmNrHjqI9", "parentPublication": { "id": "proceedings/dmdcm/2011/4413/0", "title": "Digital Media and Digital Content Management, Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/1996/7588/0/75880098", "title": "Facial Animation", "doi": null, "abstractUrl": "/proceedings-article/ca/1996/75880098/12OmNvT2oR2", "parentPublication": { "id": "proceedings/ca/1996/7588/0", "title": "Computer Animation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mmcs/1997/7819/0/00609773", "title": "Automated lip synchronisation for human-computer interaction and special effect animation", "doi": null, "abstractUrl": "/proceedings-article/mmcs/1997/00609773/12OmNvlg8pF", "parentPublication": { "id": "proceedings/mmcs/1997/7819/0", "title": "Proceedings of IEEE International Conference on Multimedia Computing and Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/acii/2015/9953/0/07344664", "title": "3D emotional facial animation synthesis with factored conditional Restricted Boltzmann Machines", "doi": null, "abstractUrl": "/proceedings-article/acii/2015/07344664/12OmNxdVh2J", "parentPublication": { "id": "proceedings/acii/2015/9953/0", "title": "2015 International Conference on Affective Computing and Intelligent Interaction (ACII)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2006/06/v1523", "title": "Expressive Facial Animation Synthesis by Learning Speech Coarticulation and Expression Spaces", "doi": null, "abstractUrl": "/journal/tg/2006/06/v1523/13rRUxASubv", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2005/03/v0341", "title": "Creating Speech-Synchronized Animation", "doi": null, "abstractUrl": "/journal/tg/2005/03/v0341/13rRUxE04tq", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ta/2022/03/09140332", "title": "Emotion Dependent Domain Adaptation for Speech Driven Affective Facial Feature Synthesis", "doi": null, "abstractUrl": "/journal/ta/2022/03/09140332/1lsnzQkrydG", "parentPublication": { "id": "trans/ta", "title": "IEEE Transactions on Affective Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNz4Bdgg", "title": "2010 Seventh International Conference on Computer Graphics, Imaging and Visualization", "acronym": "cgiv", "groupId": "1001775", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNxH9Xgx", "doi": "10.1109/CGIV.2010.11", "title": "Expressive MPEG-4 Facial Animation Using Quadratic Deformation Models", "normalizedTitle": "Expressive MPEG-4 Facial Animation Using Quadratic Deformation Models", "abstract": "In this paper we propose an approach compliant with the MPEG-4 standard to synthesize and control facial expressions generated using 3D facial models. This is achieved by establishing the MPEG-4 facial animation standard conformity with the quadratic deformation model representations of facial expressions. This conformity allows us to utilize the MPEG-4 facial animation parameters (FAPs) with the quadratic deformation tables, as a higher layer, to compute the FAP values. The FAP values for an expression E are computed by performing a linear mapping between a set of transformed MPEG-4 FAP points (using quadratic deformation models) and the 3D facial model semantics. The nature of the quadratic deformation model representations of facial expressions can be employed to synthesize and control the six main expressions (smile, sad, fear, surprise, anger, and disgust). Using Whissel's psychological studies on emotions we compute an interpolation parameter that is used to synthesize intermediate facial expressions. The paper presents results of experimental studies performed using the Greta embodied conversational agent. The achieved results are promising and can lead to future research in synthesizing a wider range of facial expressions.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper we propose an approach compliant with the MPEG-4 standard to synthesize and control facial expressions generated using 3D facial models. This is achieved by establishing the MPEG-4 facial animation standard conformity with the quadratic deformation model representations of facial expressions. This conformity allows us to utilize the MPEG-4 facial animation parameters (FAPs) with the quadratic deformation tables, as a higher layer, to compute the FAP values. The FAP values for an expression E are computed by performing a linear mapping between a set of transformed MPEG-4 FAP points (using quadratic deformation models) and the 3D facial model semantics. The nature of the quadratic deformation model representations of facial expressions can be employed to synthesize and control the six main expressions (smile, sad, fear, surprise, anger, and disgust). Using Whissel's psychological studies on emotions we compute an interpolation parameter that is used to synthesize intermediate facial expressions. The paper presents results of experimental studies performed using the Greta embodied conversational agent. The achieved results are promising and can lead to future research in synthesizing a wider range of facial expressions.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper we propose an approach compliant with the MPEG-4 standard to synthesize and control facial expressions generated using 3D facial models. This is achieved by establishing the MPEG-4 facial animation standard conformity with the quadratic deformation model representations of facial expressions. This conformity allows us to utilize the MPEG-4 facial animation parameters (FAPs) with the quadratic deformation tables, as a higher layer, to compute the FAP values. The FAP values for an expression E are computed by performing a linear mapping between a set of transformed MPEG-4 FAP points (using quadratic deformation models) and the 3D facial model semantics. The nature of the quadratic deformation model representations of facial expressions can be employed to synthesize and control the six main expressions (smile, sad, fear, surprise, anger, and disgust). Using Whissel's psychological studies on emotions we compute an interpolation parameter that is used to synthesize intermediate facial expressions. The paper presents results of experimental studies performed using the Greta embodied conversational agent. The achieved results are promising and can lead to future research in synthesizing a wider range of facial expressions.", "fno": "4166a009", "keywords": [ "MPEG 4 Facial Animation", "Quadratic Deformation Models", "Facial Animation", "Facial Expression Synthesis" ], "authors": [ { "affiliation": null, "fullName": "Mohammad Obaid", "givenName": "Mohammad", "surname": "Obaid", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ramakrishnan Mukundan", "givenName": "Ramakrishnan", "surname": "Mukundan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Mark Billinghurst", "givenName": "Mark", "surname": "Billinghurst", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Catherine Pelachaud", "givenName": "Catherine", "surname": "Pelachaud", "__typename": "ArticleAuthorType" } ], "idPrefix": "cgiv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-08-01T00:00:00", "pubType": "proceedings", "pages": "9-14", "year": "2010", "issn": null, "isbn": "978-0-7695-4166-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4166a005", "articleId": "12OmNvpNIuU", "__typename": "AdjacentArticleType" }, "next": { "fno": "4166a017", "articleId": "12OmNy314bN", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icip/1998/8821/2/882120924", "title": "Flexible face animation using MPEG-4/SNHC parameter streams", "doi": null, "abstractUrl": "/proceedings-article/icip/1998/882120924/12OmNBV9IbG", "parentPublication": { "id": "proceedings/icip/1998/8821/3", "title": "Image Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icig/2007/2929/0/29290874", "title": "Modeling Expressive Wrinkles of Face For Animation", "doi": null, "abstractUrl": "/proceedings-article/icig/2007/29290874/12OmNBV9Iif", "parentPublication": { "id": "proceedings/icig/2007/2929/0", "title": "Image and Graphics, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dicta/2009/3866/0/3866a264", "title": "A Quadratic Deformation Model for Facial Expression Recognition", "doi": null, "abstractUrl": "/proceedings-article/dicta/2009/3866a264/12OmNC4wtDl", "parentPublication": { "id": "proceedings/dicta/2009/3866/0", "title": "2009 Digital Image Computing: Techniques and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ss/2002/1552/0/15520395", "title": "Animating 3D Facial Models with MPEG-4 FaceDefTables", "doi": null, "abstractUrl": "/proceedings-article/ss/2002/15520395/12OmNrJROTd", "parentPublication": { "id": "proceedings/ss/2002/1552/0", "title": "Simulation Symposium, Annual", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/1996/7588/0/75880098", "title": "Facial Animation", "doi": null, "abstractUrl": "/proceedings-article/ca/1996/75880098/12OmNvT2oR2", "parentPublication": { "id": "proceedings/ca/1996/7588/0", "title": "Computer Animation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgiv/2009/3789/0/3789a044", "title": "Facial Expression Representation Using a Quadratic Deformation Model", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2009/3789a044/12OmNxEBz8F", "parentPublication": { "id": "proceedings/cgiv/2009/3789/0", "title": "2009 Sixth International Conference on Computer Graphics, Imaging and Visualization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/acii/2015/9953/0/07344664", "title": "3D emotional facial animation synthesis with factored conditional Restricted Boltzmann Machines", "doi": null, "abstractUrl": "/proceedings-article/acii/2015/07344664/12OmNxdVh2J", "parentPublication": { "id": "proceedings/acii/2015/9953/0", "title": "2015 International Conference on Affective Computing and Intelligent Interaction (ACII)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2004/8603/1/01394141", "title": "MPEG-4 compliant reproduction of face animation created in Maya", "doi": null, "abstractUrl": "/proceedings-article/icme/2004/01394141/12OmNy7h3bg", "parentPublication": { "id": "proceedings/icme/2004/8603/1", "title": "2004 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icassp/2004/8484/3/01326624", "title": "Inner lip feature extraction for MPEG-4 facial animation", "doi": null, "abstractUrl": "/proceedings-article/icassp/2004/01326624/12OmNyen1ka", "parentPublication": { "id": "proceedings/icassp/2004/8484/3", "title": "2004 IEEE International Conference on Acoustics, Speech, and Signal Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08960398", "title": "Data-Driven 3D Neck Modeling and Animation", "doi": null, "abstractUrl": "/journal/tg/2021/07/08960398/1gC2pML2yuk", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqG0SXH", "title": "Computer Graphics and Applications, Pacific Conference on", "acronym": "pg", "groupId": "1000130", "volume": "0", "displayVolume": "0", "year": "2000", "__typename": "ProceedingType" }, "article": { "id": "12OmNyqzLXk", "doi": "10.1109/PCCGA.2000.883959", "title": "Control of Feature-Point-Driven Facial Animation Using a Hypothetical Face", "normalizedTitle": "Control of Feature-Point-Driven Facial Animation Using a Hypothetical Face", "abstract": "A new approach for the generation of feature-point-driven facial animation is presented. This approach is based on the construction of a hypothetical face formed by connecting face feature points into a net and representing each facet of the net by a mathematically expressed surfaces to control the deformation of a real face model. Changing both the locations of the feature points and the tangents on them controls the deformation. The experiment results show that the hypothetical-surface-based method can generated almost identical facial expression as real face does.", "abstracts": [ { "abstractType": "Regular", "content": "A new approach for the generation of feature-point-driven facial animation is presented. This approach is based on the construction of a hypothetical face formed by connecting face feature points into a net and representing each facet of the net by a mathematically expressed surfaces to control the deformation of a real face model. Changing both the locations of the feature points and the tangents on them controls the deformation. The experiment results show that the hypothetical-surface-based method can generated almost identical facial expression as real face does.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "A new approach for the generation of feature-point-driven facial animation is presented. This approach is based on the construction of a hypothetical face formed by connecting face feature points into a net and representing each facet of the net by a mathematically expressed surfaces to control the deformation of a real face model. Changing both the locations of the feature points and the tangents on them controls the deformation. The experiment results show that the hypothetical-surface-based method can generated almost identical facial expression as real face does.", "fno": "08680359", "keywords": [ "Facial Animation", "Feature Point Driven", "Hypothetical Face", "Hypothetical Surface" ], "authors": [ { "affiliation": "Academia Sinica and National Taiwan University", "fullName": "Ming-Shing Su", "givenName": "Ming-Shing", "surname": "Su", "__typename": "ArticleAuthorType" }, { "affiliation": "Academia Sinica and National Taiwan University", "fullName": "Kuo-Young Cheng", "givenName": "Kuo-Young", "surname": "Cheng", "__typename": "ArticleAuthorType" }, { "affiliation": "Academia Sinica", "fullName": "Ming-Tat Ko", "givenName": "Ming-Tat", "surname": "Ko", "__typename": "ArticleAuthorType" } ], "idPrefix": "pg", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "2000-10-01T00:00:00", "pubType": "proceedings", "pages": "359", "year": "2000", "issn": null, "isbn": "0-7695-0868-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08680348", "articleId": "12OmNwoPtts", "__typename": "AdjacentArticleType" }, "next": { "fno": "08680370", "articleId": "12OmNxzuMJv", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "1cI6akLvAuQ", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1cJ0YZ9Bfgs", "doi": "10.1109/VR.2019.8798145", "title": "Speech-Driven Facial Animation by LSTM-RNN for Communication Use", "normalizedTitle": "Speech-Driven Facial Animation by LSTM-RNN for Communication Use", "abstract": "The goal of this research is developing a system that a rich facial animation can be used in communication is generated from only speech. Generally, a source of the generating facial animation is a camera. Using cameras as an input source, it causes limitations of the angle of view of the camera or problems that cannot be aware of the human face, depending on the orientation of the face. Therefore, it is reasonable for developing a system for generating a facial animation using only voice. In this study, we generate facial expressions from only speech using LSTM-RNN. Comparing 3 patterns of speech analysis data, we showed that the proposed method using A-weighting is effective for facial expression estimation.", "abstracts": [ { "abstractType": "Regular", "content": "The goal of this research is developing a system that a rich facial animation can be used in communication is generated from only speech. Generally, a source of the generating facial animation is a camera. Using cameras as an input source, it causes limitations of the angle of view of the camera or problems that cannot be aware of the human face, depending on the orientation of the face. Therefore, it is reasonable for developing a system for generating a facial animation using only voice. In this study, we generate facial expressions from only speech using LSTM-RNN. Comparing 3 patterns of speech analysis data, we showed that the proposed method using A-weighting is effective for facial expression estimation.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The goal of this research is developing a system that a rich facial animation can be used in communication is generated from only speech. Generally, a source of the generating facial animation is a camera. Using cameras as an input source, it causes limitations of the angle of view of the camera or problems that cannot be aware of the human face, depending on the orientation of the face. Therefore, it is reasonable for developing a system for generating a facial animation using only voice. In this study, we generate facial expressions from only speech using LSTM-RNN. Comparing 3 patterns of speech analysis data, we showed that the proposed method using A-weighting is effective for facial expression estimation.", "fno": "08798145", "keywords": [ "Cameras", "Computer Animation", "Face Recognition", "Recurrent Neural Nets", "Speech Processing", "Cameras", "Input Source", "Human Face", "Facial Expressions", "LSTM RNN", "Speech Analysis Data", "Facial Expression Estimation", "Speech Driven Facial Animation", "Rich Facial Animation", "Facial Animation Generation", "A Weighting", "Facial Animation", "Cameras", "Face", "Shape", "Deep Learning", "Spectrogram", "Human Centered Computing", "Visualization", "Visualization Techniques", "Treemaps", "Visualization Design And Evaluation Methods" ], "authors": [ { "affiliation": "Osaka University, Japan", "fullName": "Ryosuke Nishimura", "givenName": "Ryosuke", "surname": "Nishimura", "__typename": "ArticleAuthorType" }, { "affiliation": "Nara Institute of Science and Technology, Japan", "fullName": "Nobuchika Sakata", "givenName": "Nobuchika", "surname": "Sakata", "__typename": "ArticleAuthorType" }, { "affiliation": "Osaka University, Japan", "fullName": "Tomu Tominaga", "givenName": "Tomu", "surname": "Tominaga", "__typename": "ArticleAuthorType" }, { "affiliation": "Kwansei Gakuin University, Japan", "fullName": "Yoshinori Hijikata", "givenName": "Yoshinori", "surname": "Hijikata", "__typename": "ArticleAuthorType" }, { "affiliation": "Osaka University, Japan", "fullName": "Kensuke Harada", "givenName": "Kensuke", "surname": "Harada", "__typename": "ArticleAuthorType" }, { "affiliation": "Nara Institute of Science and Technology, Japan", "fullName": "Kiyoshi Kiyokawa", "givenName": "Kiyoshi", "surname": "Kiyokawa", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-03-01T00:00:00", "pubType": "proceedings", "pages": "1102-1103", "year": "2019", "issn": null, "isbn": "978-1-7281-1377-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08797778", "articleId": "1cJ19JavYlO", "__typename": "AdjacentArticleType" }, "next": { "fno": "08797912", "articleId": "1cJ1gTRZdIs", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccsee/2012/4647/3/4647c434", "title": "A Survey of Computer Facial Animation Techniques", "doi": null, "abstractUrl": "/proceedings-article/iccsee/2012/4647c434/12OmNAXxXhU", "parentPublication": { "id": "proceedings/iccsee/2012/4647/3", "title": "Computer Science and Electronics Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pg/2002/1784/0/17840077", "title": "\"May I talk to you? :-)\" — Facial Animation from Text", "doi": null, "abstractUrl": "/proceedings-article/pg/2002/17840077/12OmNAkWveH", "parentPublication": { "id": "proceedings/pg/2002/1784/0", "title": "Computer Graphics and Applications, Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2014/4717/0/06890554", "title": "Realtime speech-driven facial animation using Gaussian Mixture Models", "doi": null, "abstractUrl": "/proceedings-article/icmew/2014/06890554/12OmNBC8Ayh", "parentPublication": { "id": "proceedings/icmew/2014/4717/0", "title": "2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dmdcm/2011/4413/0/4413a132", "title": "Towards 3D Communications: Real Time Emotion Driven 3D Virtual Facial Animation", "doi": null, "abstractUrl": "/proceedings-article/dmdcm/2011/4413a132/12OmNrHjqI9", "parentPublication": { "id": "proceedings/dmdcm/2011/4413/0", "title": "Digital Media and Digital Content Management, Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/1996/7588/0/75880098", "title": "Facial Animation", "doi": null, "abstractUrl": "/proceedings-article/ca/1996/75880098/12OmNvT2oR2", "parentPublication": { "id": "proceedings/ca/1996/7588/0", "title": "Computer Animation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2017/0733/0/0733c328", "title": "Speech-Driven 3D Facial Animation with Implicit Emotional Awareness: A Deep Learning Approach", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2017/0733c328/12OmNxE2mG1", "parentPublication": { "id": "proceedings/cvprw/2017/0733/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2014/4761/0/06890231", "title": "Real-time control of 3D facial animation", "doi": null, "abstractUrl": "/proceedings-article/icme/2014/06890231/12OmNyOHG1A", "parentPublication": { "id": "proceedings/icme/2014/4761/0", "title": "2014 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2006/06/v1523", "title": "Expressive Facial Animation Synthesis by Learning Speech Coarticulation and Expression Spaces", "doi": null, "abstractUrl": "/journal/tg/2006/06/v1523/13rRUxASubv", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2005/03/v0341", "title": "Creating Speech-Synchronized Animation", "doi": null, "abstractUrl": "/journal/tg/2005/03/v0341/13rRUxE04tq", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/12/09524465", "title": "Geometry-Guided Dense Perspective Network for Speech-Driven Facial Animation", "doi": null, "abstractUrl": "/journal/tg/2022/12/09524465/1wpqCsqBU6Q", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1B12DGrwoyQ", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "acronym": "wacv", "groupId": "1000040", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1B12RNuAqvS", "doi": "10.1109/WACV51458.2022.00070", "title": "Shadow Art Revisited: A Differentiable Rendering Based Approach", "normalizedTitle": "Shadow Art Revisited: A Differentiable Rendering Based Approach", "abstract": "While recent learning-based methods have been observed to be superior for several vision-related applications, their potential in generating artistic effects has not been explored much. One such exciting application is Shadow Art - a unique form of sculptural art that produces artistic effects through 2D shadows cast by a 3D sculpture. In this work, we revisit shadow art using differentiable rendering-based optimization frameworks to obtain the 3D sculpture from a set of shadow (binary) images and their corresponding projection information. Specifically, we discuss shape optimization through voxel as well as mesh-based differentiable renderers. Our choice of using differentiable rendering for generating shadow art sculptures can be attributed to its ability to learn the underlying 3D geometry solely from image data, thus reducing the dependence on 3D ground truth. The qualitative and quantitative results demonstrate the potential of the proposed framework in generating complex 3D sculptures that transcend the ones seen in contemporary art pieces using just a set of shadow images as input. Further, we demonstrate the generation of 3D sculptures to cast shadows of faces, animated movie characters, and the applicability of the proposed framework to sketch-based 3D reconstruction of the underlying shapes.", "abstracts": [ { "abstractType": "Regular", "content": "While recent learning-based methods have been observed to be superior for several vision-related applications, their potential in generating artistic effects has not been explored much. One such exciting application is Shadow Art - a unique form of sculptural art that produces artistic effects through 2D shadows cast by a 3D sculpture. In this work, we revisit shadow art using differentiable rendering-based optimization frameworks to obtain the 3D sculpture from a set of shadow (binary) images and their corresponding projection information. Specifically, we discuss shape optimization through voxel as well as mesh-based differentiable renderers. Our choice of using differentiable rendering for generating shadow art sculptures can be attributed to its ability to learn the underlying 3D geometry solely from image data, thus reducing the dependence on 3D ground truth. The qualitative and quantitative results demonstrate the potential of the proposed framework in generating complex 3D sculptures that transcend the ones seen in contemporary art pieces using just a set of shadow images as input. Further, we demonstrate the generation of 3D sculptures to cast shadows of faces, animated movie characters, and the applicability of the proposed framework to sketch-based 3D reconstruction of the underlying shapes.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "While recent learning-based methods have been observed to be superior for several vision-related applications, their potential in generating artistic effects has not been explored much. One such exciting application is Shadow Art - a unique form of sculptural art that produces artistic effects through 2D shadows cast by a 3D sculpture. In this work, we revisit shadow art using differentiable rendering-based optimization frameworks to obtain the 3D sculpture from a set of shadow (binary) images and their corresponding projection information. Specifically, we discuss shape optimization through voxel as well as mesh-based differentiable renderers. Our choice of using differentiable rendering for generating shadow art sculptures can be attributed to its ability to learn the underlying 3D geometry solely from image data, thus reducing the dependence on 3D ground truth. The qualitative and quantitative results demonstrate the potential of the proposed framework in generating complex 3D sculptures that transcend the ones seen in contemporary art pieces using just a set of shadow images as input. Further, we demonstrate the generation of 3D sculptures to cast shadows of faces, animated movie characters, and the applicability of the proposed framework to sketch-based 3D reconstruction of the underlying shapes.", "fno": "091500a628", "keywords": [ "Art", "Computational Geometry", "Computer Animation", "Computer Vision", "Image Reconstruction", "Learning Artificial Intelligence", "Optimisation", "Rendering Computer Graphics", "Solid Modelling", "Shadow Art", "Learning Based Methods", "Vision Related Applications", "Artistic Effects", "Sculptural Art", "Differentiable Rendering Based Optimization Frameworks", "Shadow Images", "Projection Information", "Shape Optimization", "Mesh Based Differentiable Renderers", "Shadow Art Sculptures", "Underlying 3 D Geometry", "3 D Ground Truth", "Complex 3 D Sculptures", "Contemporary Art Pieces", "3 D Reconstruction", "Learning Systems", "Three Dimensional Displays", "Art", "Shape", "Pipelines", "Rendering Computer Graphics", "Motion Pictures", "3 D Computer Vision Vision For Graphics" ], "authors": [ { "affiliation": "IIT Gandhinagar,CVIG Lab", "fullName": "Kaustubh Sadekar", "givenName": "Kaustubh", "surname": "Sadekar", "__typename": "ArticleAuthorType" }, { "affiliation": "IIT Gandhinagar,CVIG Lab", "fullName": "Ashish Tiwari", "givenName": "Ashish", "surname": "Tiwari", "__typename": "ArticleAuthorType" }, { "affiliation": "IIT Gandhinagar,CVIG Lab", "fullName": "Shanmuganathan Raman", "givenName": "Shanmuganathan", "surname": "Raman", "__typename": "ArticleAuthorType" } ], "idPrefix": "wacv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-01-01T00:00:00", "pubType": "proceedings", "pages": "628-636", "year": "2022", "issn": null, "isbn": "978-1-6654-0915-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "091500a617", "articleId": "1B13liUsWVG", "__typename": "AdjacentArticleType" }, "next": { "fno": "091500a637", "articleId": "1B13sjY8Yms", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/nicoint/2017/5332/0/5332a080", "title": "STAT (U) ES: An Interactive Community Engaged Art Using Projection Mapping and Facial Recognition System", "doi": null, "abstractUrl": "/proceedings-article/nicoint/2017/5332a080/12OmNrYlmF0", "parentPublication": { "id": "proceedings/nicoint/2017/5332/0", "title": "2017 Nicograph International (NicoInt)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2008/01/ttg2008010135", "title": "Clip Art Rendering of Smooth Isosurfaces", "doi": null, "abstractUrl": "/journal/tg/2008/01/ttg2008010135/13rRUyYBlgs", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200g068", "title": "Differentiable Surface Rendering via Non-Differentiable Sampling", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200g068/1BmFpmQFMKA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200n3087", "title": "Efficient and Differentiable Shadow Computation for Inverse Problems", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200n3087/1BmFvUmmGGY", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600i625", "title": "Differentiable Stereopsis: Meshes from multiple views using differentiable rendering", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600i625/1H0NABVhMdO", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800m2221", "title": "Autolabeling 3D Objects With Differentiable Rendering of SDF Shape Priors", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800m2221/1m3nWEcnnzy", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800b248", "title": "SDFDiff: Differentiable Rendering of Signed Distance Fields for 3D Shape Optimization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800b248/1m3nwOmON4Q", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800d501", "title": "Differentiable Volumetric Rendering: Learning Implicit 3D Representations Without 3D Supervision", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800d501/1m3nwXQXEAw", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2021/1838/0/255600a170", "title": "Bidirectional Shadow Rendering for Interactive Mixed 360° Videos", "doi": null, "abstractUrl": "/proceedings-article/vr/2021/255600a170/1tuAEjkRUZy", "parentPublication": { "id": "proceedings/vr/2021/1838/0", "title": "2021 IEEE Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/01/09552224", "title": "Differentiable Direct Volume Rendering", "doi": null, "abstractUrl": "/journal/tg/2022/01/09552224/1xibZvRmYzm", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1BmEezmpGrm", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1BmLryCiwjm", "doi": "10.1109/ICCV48922.2021.01106", "title": "Learning to Regress Bodies from Images using Differentiable Semantic Rendering", "normalizedTitle": "Learning to Regress Bodies from Images using Differentiable Semantic Rendering", "abstract": "Learning to regress 3D human body shape and pose (e.g. SMPL parameters) from monocular images typically exploits losses on 2D keypoints, silhouettes, and/or part-segmentation when 3D training data is not available. Such losses, however, are limited because 2D keypoints do not supervise body shape and segmentations of people in clothing do not match projected minimally-clothed SMPL shapes. To exploit richer image information about clothed people, we introduce higher-level semantic information about clothing to penalize clothed and non-clothed regions of the human body differently. To do so, we train a body regressor using a novel \"Differentiable Semantic Rendering (DSR)\" loss. For Minimally-Clothed (MC) regions, we define the DSR-MC loss, which encourages a tight match between a rendered SMPL body and the minimally-clothed regions of the image. For clothed regions, we define the DSR-C loss to encourage the rendered SMPL body to be inside the clothing mask. To ensure end-to-end differentiable training, we learn a semantic clothing prior for SMPL vertices from thousands of clothed human scans. We perform extensive qualitative and quantitative experiments to evaluate the role of clothing semantics on the accuracy of 3D human pose and shape estimation. We outperform all previous state-of-the-art methods on 3DPW and Human3.6M and obtain on par results on MPI-INF-3DHP. Code and trained models are available for research at https://dsr.is.tue.mpg.de/.", "abstracts": [ { "abstractType": "Regular", "content": "Learning to regress 3D human body shape and pose (e.g. SMPL parameters) from monocular images typically exploits losses on 2D keypoints, silhouettes, and/or part-segmentation when 3D training data is not available. Such losses, however, are limited because 2D keypoints do not supervise body shape and segmentations of people in clothing do not match projected minimally-clothed SMPL shapes. To exploit richer image information about clothed people, we introduce higher-level semantic information about clothing to penalize clothed and non-clothed regions of the human body differently. To do so, we train a body regressor using a novel \"Differentiable Semantic Rendering (DSR)\" loss. For Minimally-Clothed (MC) regions, we define the DSR-MC loss, which encourages a tight match between a rendered SMPL body and the minimally-clothed regions of the image. For clothed regions, we define the DSR-C loss to encourage the rendered SMPL body to be inside the clothing mask. To ensure end-to-end differentiable training, we learn a semantic clothing prior for SMPL vertices from thousands of clothed human scans. We perform extensive qualitative and quantitative experiments to evaluate the role of clothing semantics on the accuracy of 3D human pose and shape estimation. We outperform all previous state-of-the-art methods on 3DPW and Human3.6M and obtain on par results on MPI-INF-3DHP. Code and trained models are available for research at https://dsr.is.tue.mpg.de/.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Learning to regress 3D human body shape and pose (e.g. SMPL parameters) from monocular images typically exploits losses on 2D keypoints, silhouettes, and/or part-segmentation when 3D training data is not available. Such losses, however, are limited because 2D keypoints do not supervise body shape and segmentations of people in clothing do not match projected minimally-clothed SMPL shapes. To exploit richer image information about clothed people, we introduce higher-level semantic information about clothing to penalize clothed and non-clothed regions of the human body differently. To do so, we train a body regressor using a novel \"Differentiable Semantic Rendering (DSR)\" loss. For Minimally-Clothed (MC) regions, we define the DSR-MC loss, which encourages a tight match between a rendered SMPL body and the minimally-clothed regions of the image. For clothed regions, we define the DSR-C loss to encourage the rendered SMPL body to be inside the clothing mask. To ensure end-to-end differentiable training, we learn a semantic clothing prior for SMPL vertices from thousands of clothed human scans. We perform extensive qualitative and quantitative experiments to evaluate the role of clothing semantics on the accuracy of 3D human pose and shape estimation. We outperform all previous state-of-the-art methods on 3DPW and Human3.6M and obtain on par results on MPI-INF-3DHP. Code and trained models are available for research at https://dsr.is.tue.mpg.de/.", "fno": "281200l1230", "keywords": [ "Training", "Solid Modeling", "Three Dimensional Displays", "Shape", "Semantics", "Clothing", "Estimation", "Gestures And Body Pose", "3 D From A Single Image And Shape From X" ], "authors": [ { "affiliation": "Max Planck Institute for Intelligent Systems,Tübingen,Germany", "fullName": "Sai Kumar Dwivedi", "givenName": "Sai Kumar", "surname": "Dwivedi", "__typename": "ArticleAuthorType" }, { "affiliation": "Max Planck Institute for Intelligent Systems,Tübingen,Germany", "fullName": "Nikos Athanasiou", "givenName": "Nikos", "surname": "Athanasiou", "__typename": "ArticleAuthorType" }, { "affiliation": "Max Planck Institute for Intelligent Systems,Tübingen,Germany", "fullName": "Muhammed Kocabas", "givenName": "Muhammed", "surname": "Kocabas", "__typename": "ArticleAuthorType" }, { "affiliation": "Max Planck Institute for Intelligent Systems,Tübingen,Germany", "fullName": "Michael J. Black", "givenName": "Michael J.", "surname": "Black", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-10-01T00:00:00", "pubType": "proceedings", "pages": "11230-11239", "year": "2021", "issn": null, "isbn": "978-1-6654-2812-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "281200l1220", "articleId": "1BmKOT20mJ2", "__typename": "AdjacentArticleType" }, "next": { "fno": "281200l1240", "articleId": "1BmL6BWkQIE", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2021/2812/0/281200k0954", "title": "The Power of Points for Modeling Humans in Clothing", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200k0954/1BmLrmWbNuM", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2022/5670/0/567000a679", "title": "Neural Point-based Shape Modeling of Humans in Challenging Clothing", "doi": null, "abstractUrl": "/proceedings-article/3dv/2022/567000a679/1KYsvi8qLS0", "parentPublication": { "id": "proceedings/3dv/2022/5670/0", "title": "2022 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2019/3131/0/313100a643", "title": "360-Degree Textures of People in Clothing from a Single Image", "doi": null, "abstractUrl": "/proceedings-article/3dv/2019/313100a643/1ezRDPNhkpW", "parentPublication": { "id": "proceedings/3dv/2019/3131/0", "title": "2019 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300k0510", "title": "VTNFP: An Image-Based Virtual Try-On Network With Body and Clothing Feature Preservation", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300k0510/1hVlSD4rLA4", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300e351", "title": "Shape-Aware Human Pose and Shape Reconstruction Using Multi-View Images", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300e351/1hVlr4uihb2", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300f419", "title": "Multi-Garment Net: Learning to Dress 3D People From Images", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300f419/1hVlwZpXtZK", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800g468", "title": "Learning to Dress 3D People in Generative Clothing", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800g468/1m3nwUHFD68", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900q6077", "title": "SCALE: Modeling Clothed Humans with a Surface Codec of Articulated Local Elements", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900q6077/1yeIj9jyjks", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900c885", "title": "SCANimate: Weakly Supervised Learning of Skinned Clothed Avatar Networks", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900c885/1yeJrwzKCyI", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900n3463", "title": "AGORA: Avatars in Geography Optimized for Regression Analysis", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900n3463/1yeKylgDVa8", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1IlNS8T3xtu", "title": "2022 5th International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE)", "acronym": "aemcse", "groupId": "9948218", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1IlObcruRxK", "doi": "10.1109/AEMCSE55572.2022.00041", "title": "3D Communication System Integrating 3D Reconstruction and Rendering Display", "normalizedTitle": "3D Communication System Integrating 3D Reconstruction and Rendering Display", "abstract": "In order to solve the problem that the existing 2D video communication has a high bandwidth occupancy rate, the inability to present three-dimensional portraits of characters and the high dependence of three-dimensional projection technology on hardware, we propose a 3D communication framework that integrates 3D reconstruction and rendering display. The framework transmits the video stream to the video encoding module. The video coding module mainly completes the work of face detection, key point feature extraction and depth map feature coding. It is transmitted to the terminal device through the transmission protocol for video decoding and reconstruction. Then combined with the 3D rendering algorithm, virtual viewpoint synthesis and stereo image encoding are performed to realize the 3D reconstruction of the real face. The frame structure opens up the technical path from ordinary 2D video acquisition to 3D communication system implementation. Experiments show that the use of this communication framework for real face 3D reconstruction has higher accuracy and better information fusion. The reconstruction rate can reach 15fps.", "abstracts": [ { "abstractType": "Regular", "content": "In order to solve the problem that the existing 2D video communication has a high bandwidth occupancy rate, the inability to present three-dimensional portraits of characters and the high dependence of three-dimensional projection technology on hardware, we propose a 3D communication framework that integrates 3D reconstruction and rendering display. The framework transmits the video stream to the video encoding module. The video coding module mainly completes the work of face detection, key point feature extraction and depth map feature coding. It is transmitted to the terminal device through the transmission protocol for video decoding and reconstruction. Then combined with the 3D rendering algorithm, virtual viewpoint synthesis and stereo image encoding are performed to realize the 3D reconstruction of the real face. The frame structure opens up the technical path from ordinary 2D video acquisition to 3D communication system implementation. Experiments show that the use of this communication framework for real face 3D reconstruction has higher accuracy and better information fusion. The reconstruction rate can reach 15fps.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In order to solve the problem that the existing 2D video communication has a high bandwidth occupancy rate, the inability to present three-dimensional portraits of characters and the high dependence of three-dimensional projection technology on hardware, we propose a 3D communication framework that integrates 3D reconstruction and rendering display. The framework transmits the video stream to the video encoding module. The video coding module mainly completes the work of face detection, key point feature extraction and depth map feature coding. It is transmitted to the terminal device through the transmission protocol for video decoding and reconstruction. Then combined with the 3D rendering algorithm, virtual viewpoint synthesis and stereo image encoding are performed to realize the 3D reconstruction of the real face. The frame structure opens up the technical path from ordinary 2D video acquisition to 3D communication system implementation. Experiments show that the use of this communication framework for real face 3D reconstruction has higher accuracy and better information fusion. The reconstruction rate can reach 15fps.", "fno": "847400a167", "keywords": [ "Decoding", "Feature Extraction", "Image Reconstruction", "Rendering Computer Graphics", "Stereo Image Processing", "Three Dimensional Displays", "Video Coding", "Video Communication", "Video Streaming", "3 D Communication System Implementation", "3 D Communication System Integrating 3 D Reconstruction", "3 D Rendering Algorithm", "Communication Framework", "Depth Map Feature Coding", "Existing 2 D Video Communication", "Face 3 D Reconstruction", "High Bandwidth Occupancy Rate", "High Dependence", "Key Point Feature Extraction", "Ordinary 2 D Video Acquisition", "Reconstruction Rate", "Rendering Display", "Three Dimensional Projection Technology", "Video Coding Module", "Video Decoding", "Video Encoding Module", "Video Stream", "Computers", "Video Coding", "Three Dimensional Displays", "Two Dimensional Displays", "Streaming Media", "Rendering Computer Graphics", "Feature Extraction", "Computer Vision", "3 D Communication", "3 D Rendering" ], "authors": [ { "affiliation": "China Telecom Corporation Limited,Beijing Research Institute,Beijing,China,102209", "fullName": "Chaoying Zhang", "givenName": "Chaoying", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": "China Telecom Corporation Limited,Beijing Research Institute,Beijing,China,102209", "fullName": "Minglan Su", "givenName": "Minglan", "surname": "Su", "__typename": "ArticleAuthorType" }, { "affiliation": "China Telecom Corporation Limited,Beijing Research Institute,Beijing,China,102209", "fullName": "Qiaoqiao Liu", "givenName": "Qiaoqiao", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "China Telecom Corporation Limited,Beijing Research Institute,Beijing,China,102209", "fullName": "Mingchuan Yang", "givenName": "Mingchuan", "surname": "Yang", "__typename": "ArticleAuthorType" } ], "idPrefix": "aemcse", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-04-01T00:00:00", "pubType": "proceedings", "pages": "167-170", "year": "2022", "issn": null, "isbn": "978-1-6654-8474-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "847400a163", "articleId": "1IlOh584CCQ", "__typename": "AdjacentArticleType" }, "next": { "fno": "847400a171", "articleId": "1IlNTV6OnYI", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2017/0457/0/0457b503", "title": "End-to-End 3D Face Reconstruction with Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457b503/12OmNrEL2xy", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fg/2018/2335/0/233501a780", "title": "Evaluation of Dense 3D Reconstruction from 2D Face Images in the Wild", "doi": null, "abstractUrl": "/proceedings-article/fg/2018/233501a780/12OmNyPQ4HL", "parentPublication": { "id": "proceedings/fg/2018/2335/0", "title": "2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2014/4258/0/4258a001", "title": "3D Face Reconstruction from Video Using 3D Morphable Model and Silhouette", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2014/4258a001/12OmNz5JBYT", "parentPublication": { "id": "proceedings/sibgrapi/2014/4258/0", "title": "2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2020/03/08571265", "title": "Joint Face Alignment and 3D Face Reconstruction with Application to Face Recognition", "doi": null, "abstractUrl": "/journal/tp/2020/03/08571265/17D45WnnFYh", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000f216", "title": "Disentangling Features in 3D Face Shapes for Joint Face Reconstruction and Recognition", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000f216/17D45WrVgfL", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aicis/2018/9188/0/918800a259", "title": "Overview: 3D Video from Capture to Display", "doi": null, "abstractUrl": "/proceedings-article/aicis/2018/918800a259/17PYEm2zquf", "parentPublication": { "id": "proceedings/aicis/2018/9188/0", "title": "2018 1st Annual International Conference on Information and Sciences (AiCIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csci/2021/5841/0/584100b611", "title": "3D Face Reconstruction from Front and Profile Images for Low Computational Devices", "doi": null, "abstractUrl": "/proceedings-article/csci/2021/584100b611/1EpLDQz7nZC", "parentPublication": { "id": "proceedings/csci/2021/5841/0", "title": "2021 International Conference on Computational Science and Computational Intelligence (CSCI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aemcse/2022/8474/0/847400a854", "title": "Research on Holographic Communication System Based on 3D Face Reconstruction", "doi": null, "abstractUrl": "/proceedings-article/aemcse/2022/847400a854/1IlNWfcNhTi", "parentPublication": { "id": "proceedings/aemcse/2022/8474/0", "title": "2022 5th International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2019/5023/0/502300d082", "title": "The 2nd 3D Face Alignment in the Wild Challenge (3DFAW-Video): Dense Reconstruction From Video", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2019/502300d082/1i5muuh7S6s", "parentPublication": { "id": "proceedings/iccvw/2019/5023/0", "title": "2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800d501", "title": "Differentiable Volumetric Rendering: Learning Implicit 3D Representations Without 3D Supervision", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800d501/1m3nwXQXEAw", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1hQqfuoOyHu", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1hVlfIgUyLm", "doi": "10.1109/ICCV.2019.00780", "title": "Soft Rasterizer: A Differentiable Renderer for Image-Based 3D Reasoning", "normalizedTitle": "Soft Rasterizer: A Differentiable Renderer for Image-Based 3D Reasoning", "abstract": "Rendering bridges the gap between 2D vision and 3D scenes by simulating the physical process of image formation. By inverting such renderer, one can think of a learning approach to infer 3D information from 2D images. However, standard graphics renderers involve a fundamental discretization step called rasterization, which prevents the rendering process to be differentiable, hence able to be learned. Unlike the state-of-the-art differentiable renderers, which only approximate the rendering gradient in the back propagation, we propose a truly differentiable rendering framework that is able to (1) directly render colorized mesh using differentiable functions and (2) back-propagate efficient supervision signals to mesh vertices and their attributes from various forms of image representations, including silhouette, shading and color images. The key to our framework is a novel formulation that views rendering as an aggregation function that fuses the probabilistic contributions of all mesh triangles with respect to the rendered pixels. Such formulation enables our framework to flow gradients to the occluded and far-range vertices, which cannot be achieved by the previous state-of-the-arts. We show that by using the proposed renderer, one can achieve significant improvement in 3D unsupervised single-view reconstruction both qualitatively and quantitatively. Experiments also demonstrate that our approach is able to handle the challenging tasks in image-based shape fitting, which remain nontrivial to existing differentiable renderers. Code is available at https://github.com/ShichenLiu/SoftRas.", "abstracts": [ { "abstractType": "Regular", "content": "Rendering bridges the gap between 2D vision and 3D scenes by simulating the physical process of image formation. By inverting such renderer, one can think of a learning approach to infer 3D information from 2D images. However, standard graphics renderers involve a fundamental discretization step called rasterization, which prevents the rendering process to be differentiable, hence able to be learned. Unlike the state-of-the-art differentiable renderers, which only approximate the rendering gradient in the back propagation, we propose a truly differentiable rendering framework that is able to (1) directly render colorized mesh using differentiable functions and (2) back-propagate efficient supervision signals to mesh vertices and their attributes from various forms of image representations, including silhouette, shading and color images. The key to our framework is a novel formulation that views rendering as an aggregation function that fuses the probabilistic contributions of all mesh triangles with respect to the rendered pixels. Such formulation enables our framework to flow gradients to the occluded and far-range vertices, which cannot be achieved by the previous state-of-the-arts. We show that by using the proposed renderer, one can achieve significant improvement in 3D unsupervised single-view reconstruction both qualitatively and quantitatively. Experiments also demonstrate that our approach is able to handle the challenging tasks in image-based shape fitting, which remain nontrivial to existing differentiable renderers. Code is available at https://github.com/ShichenLiu/SoftRas.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Rendering bridges the gap between 2D vision and 3D scenes by simulating the physical process of image formation. By inverting such renderer, one can think of a learning approach to infer 3D information from 2D images. However, standard graphics renderers involve a fundamental discretization step called rasterization, which prevents the rendering process to be differentiable, hence able to be learned. Unlike the state-of-the-art differentiable renderers, which only approximate the rendering gradient in the back propagation, we propose a truly differentiable rendering framework that is able to (1) directly render colorized mesh using differentiable functions and (2) back-propagate efficient supervision signals to mesh vertices and their attributes from various forms of image representations, including silhouette, shading and color images. The key to our framework is a novel formulation that views rendering as an aggregation function that fuses the probabilistic contributions of all mesh triangles with respect to the rendered pixels. Such formulation enables our framework to flow gradients to the occluded and far-range vertices, which cannot be achieved by the previous state-of-the-arts. We show that by using the proposed renderer, one can achieve significant improvement in 3D unsupervised single-view reconstruction both qualitatively and quantitatively. Experiments also demonstrate that our approach is able to handle the challenging tasks in image-based shape fitting, which remain nontrivial to existing differentiable renderers. Code is available at https://github.com/ShichenLiu/SoftRas.", "fno": "480300h707", "keywords": [ "Image Colour Analysis", "Image Reconstruction", "Image Representation", "Inference Mechanisms", "Learning Artificial Intelligence", "Rendering Computer Graphics", "Rendering Gradient", "Image Representations", "Color Images", "Rendered Pixels", "Image Based Shape Fitting", "Soft Rasterizer", "Image Based 3 D Reasoning", "Image Formation", "Learning Approach", "Standard Graphics Renderers", "Fundamental Discretization Step", "Differentiable Rendering Framework", "2 D Vision", "3 D Scenes", "Three Dimensional Displays", "Rendering Computer Graphics", "Two Dimensional Displays", "Cognition", "Standards", "Image Reconstruction", "Task Analysis" ], "authors": [ { "affiliation": "University of Southern California", "fullName": "Shichen Liu", "givenName": "Shichen", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "USC Institute for Creative Technology", "fullName": "Weikai Chen", "givenName": "Weikai", "surname": "Chen", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Southern California", "fullName": "Tianye Li", "givenName": "Tianye", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "Pinscreen/University of Southern California/USC ICT", "fullName": "Hao Li", "givenName": "Hao", "surname": "Li", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-10-01T00:00:00", "pubType": "proceedings", "pages": "7707-7716", "year": "2019", "issn": null, "isbn": "978-1-7281-4803-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "480300h697", "articleId": "1hQqyXwRbm8", "__typename": "AdjacentArticleType" }, "next": { "fno": "480300h717", "articleId": "1hVlLjR0Chq", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2018/6420/0/642000d907", "title": "Neural 3D Mesh Renderer", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000d907/17D45WHONlu", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500a628", "title": "Shadow Art Revisited: A Differentiable Rendering Based Approach", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500a628/1B12RNuAqvS", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200g068", "title": "Differentiable Surface Rendering via Non-Differentiable Sampling", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200g068/1BmFpmQFMKA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600p5284", "title": "DTA: Physical Camouflage Attacks using Differentiable Transformation Network", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600p5284/1H0KzLSKj3G", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600d992", "title": "GenDR: A Generalized Differentiable Renderer", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600d992/1H1hFjzHeDu", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600k0021", "title": "Deep 3D-to-2D Watermarking: Embedding Messages in 3D Meshes and Extracting Them from 2D Renderings", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600k0021/1H1iULNZ0nC", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2022/9062/0/09956528", "title": "Auto-augmentation with Differentiable Renderer for High-frequency Shape Recovery", "doi": null, "abstractUrl": "/proceedings-article/icpr/2022/09956528/1IHpbsFDW48", "parentPublication": { "id": "proceedings/icpr/2022/9062/0", "title": "2022 26th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08966278", "title": "A Non-Linear Differentiable CNN-Rendering Module for 3D Data Enhancement", "doi": null, "abstractUrl": "/journal/tg/2021/07/08966278/1gNEBsadHP2", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/01/09134794", "title": "A General Differentiable Mesh Renderer for Image-Based 3D Reasoning", "doi": null, "abstractUrl": "/journal/tp/2022/01/09134794/1lgLr0OBt9C", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800d501", "title": "Differentiable Volumetric Rendering: Learning Implicit 3D Representations Without 3D Supervision", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800d501/1m3nwXQXEAw", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1m3n9N02qgE", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1m3nuEI1jJm", "doi": "10.1109/CVPR42600.2020.00209", "title": "DIST: Rendering Deep Implicit Signed Distance Function With Differentiable Sphere Tracing", "normalizedTitle": "DIST: Rendering Deep Implicit Signed Distance Function With Differentiable Sphere Tracing", "abstract": "We propose a differentiable sphere tracing algorithm to bridge the gap between inverse graphics methods and the recently proposed deep learning based implicit signed distance function. Due to the nature of the implicit function, the rendering process requires tremendous function queries, which is particularly problematic when the function is represented as a neural network. We optimize both the forward and backward pass of our rendering layer to make it run efficiently with affordable memory consumption on a commodity graphics card. Our rendering method is fully differentiable such that losses can be directly computed on the rendered 2D observations, and the gradients can be propagated backward to optimize the 3D geometry. We show that our rendering method can effectively reconstruct accurate 3D shapes from various inputs, such as sparse depth and multi-view images, through inverse optimization. With the geometry based reasoning, our 3D shape prediction methods show excellent generalization capability and robustness against various noises.", "abstracts": [ { "abstractType": "Regular", "content": "We propose a differentiable sphere tracing algorithm to bridge the gap between inverse graphics methods and the recently proposed deep learning based implicit signed distance function. Due to the nature of the implicit function, the rendering process requires tremendous function queries, which is particularly problematic when the function is represented as a neural network. We optimize both the forward and backward pass of our rendering layer to make it run efficiently with affordable memory consumption on a commodity graphics card. Our rendering method is fully differentiable such that losses can be directly computed on the rendered 2D observations, and the gradients can be propagated backward to optimize the 3D geometry. We show that our rendering method can effectively reconstruct accurate 3D shapes from various inputs, such as sparse depth and multi-view images, through inverse optimization. With the geometry based reasoning, our 3D shape prediction methods show excellent generalization capability and robustness against various noises.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose a differentiable sphere tracing algorithm to bridge the gap between inverse graphics methods and the recently proposed deep learning based implicit signed distance function. Due to the nature of the implicit function, the rendering process requires tremendous function queries, which is particularly problematic when the function is represented as a neural network. We optimize both the forward and backward pass of our rendering layer to make it run efficiently with affordable memory consumption on a commodity graphics card. Our rendering method is fully differentiable such that losses can be directly computed on the rendered 2D observations, and the gradients can be propagated backward to optimize the 3D geometry. We show that our rendering method can effectively reconstruct accurate 3D shapes from various inputs, such as sparse depth and multi-view images, through inverse optimization. With the geometry based reasoning, our 3D shape prediction methods show excellent generalization capability and robustness against various noises.", "fno": "716800c016", "keywords": [ "Learning Artificial Intelligence", "Neural Nets", "Optimisation", "Rendering Computer Graphics", "Solid Modelling", "Commodity Graphics Card", "Rendering Method", "2 D Observations", "Inverse Optimization", "Geometry Based Reasoning", "3 D Shape Prediction Methods", "Deep Implicit Signed Distance Function", "Differentiable Sphere Tracing", "Inverse Graphics Methods", "Deep Learning", "Three Dimensional Displays", "Shape", "Rendering Computer Graphics", "Machine Learning", "Neural Networks", "Two Dimensional Displays", "Geometry" ], "authors": [ { "affiliation": "ETH Zurich; Tsinghua University", "fullName": "Shaohui Liu", "givenName": "Shaohui", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "Google", "fullName": "Yinda Zhang", "givenName": "Yinda", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": "ETH Zurich; Max Planck ETH Center for Learing Systems", "fullName": "Songyou Peng", "givenName": "Songyou", "surname": "Peng", "__typename": "ArticleAuthorType" }, { "affiliation": "Peking University; Peng Cheng Laboratory", "fullName": "Boxin Shi", "givenName": "Boxin", "surname": "Shi", "__typename": "ArticleAuthorType" }, { "affiliation": "ETH Zurich; Microsoft; Max Planck ETH Center for Learing Systems", "fullName": "Marc Pollefeys", "givenName": "Marc", "surname": "Pollefeys", "__typename": "ArticleAuthorType" }, { "affiliation": "ETH Zurich", "fullName": "Zhaopeng Cui", "givenName": "Zhaopeng", "surname": "Cui", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-06-01T00:00:00", "pubType": "proceedings", "pages": "2016-2025", "year": "2020", "issn": null, "isbn": "978-1-7281-7168-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "716800c006", "articleId": "1m3nI9rwo6s", "__typename": "AdjacentArticleType" }, "next": { "fno": "716800c026", "articleId": "1m3nOsO6M7u", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "trans/tg/2023/06/09705143", "title": "Adaptive Joint Optimization for 3D Reconstruction With Differentiable Rendering", "doi": null, "abstractUrl": "/journal/tg/2023/06/09705143/1AIIcwNiqxq", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500a628", "title": "Shadow Art Revisited: A Differentiable Rendering Based Approach", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500a628/1B12RNuAqvS", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200g068", "title": "Differentiable Surface Rendering via Non-Differentiable Sampling", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200g068/1BmFpmQFMKA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600d992", "title": "GenDR: A Generalized Differentiable Renderer", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600d992/1H1hFjzHeDu", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08966278", "title": "A Non-Linear Differentiable CNN-Rendering Module for 3D Data Enhancement", "doi": null, "abstractUrl": "/journal/tg/2021/07/08966278/1gNEBsadHP2", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300h707", "title": "Soft Rasterizer: A Differentiable Renderer for Image-Based 3D Reasoning", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300h707/1hVlfIgUyLm", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/01/09134794", "title": "A General Differentiable Mesh Renderer for Image-Based 3D Reasoning", "doi": null, "abstractUrl": "/journal/tp/2022/01/09134794/1lgLr0OBt9C", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800b248", "title": "SDFDiff: Differentiable Rendering of Signed Distance Fields for 3D Shape Optimization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800b248/1m3nwOmON4Q", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800d501", "title": "Differentiable Volumetric Rendering: Learning Implicit 3D Representations Without 3D Supervision", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800d501/1m3nwXQXEAw", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900b440", "title": "Pulsar: Efficient Sphere-based Neural Rendering", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900b440/1yeJSYxaxKo", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1m3n9N02qgE", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1m3nwXQXEAw", "doi": "10.1109/CVPR42600.2020.00356", "title": "Differentiable Volumetric Rendering: Learning Implicit 3D Representations Without 3D Supervision", "normalizedTitle": "Differentiable Volumetric Rendering: Learning Implicit 3D Representations Without 3D Supervision", "abstract": "Learning-based 3D reconstruction methods have shown impressive results. However, most methods require 3D supervision which is often hard to obtain for real-world datasets. Recently, several works have proposed differentiable rendering techniques to train reconstruction models from RGB images. Unfortunately, these approaches are currently restricted to voxel- and mesh-based representations, suffering from discretization or low resolution. In this work, we propose a differentiable rendering formulation for implicit shape and texture representations. Implicit representations have recently gained popularity as they represent shape and texture continuously. Our key insight is that depth gradients can be derived analytically using the concept of implicit differentiation. This allows us to learn implicit shape and texture representations directly from RGB images. We experimentally show that our single-view reconstructions rival those learned with full 3D supervision. Moreover, we find that our method can be used for multi-view 3D reconstruction, directly resulting in watertight meshes.", "abstracts": [ { "abstractType": "Regular", "content": "Learning-based 3D reconstruction methods have shown impressive results. However, most methods require 3D supervision which is often hard to obtain for real-world datasets. Recently, several works have proposed differentiable rendering techniques to train reconstruction models from RGB images. Unfortunately, these approaches are currently restricted to voxel- and mesh-based representations, suffering from discretization or low resolution. In this work, we propose a differentiable rendering formulation for implicit shape and texture representations. Implicit representations have recently gained popularity as they represent shape and texture continuously. Our key insight is that depth gradients can be derived analytically using the concept of implicit differentiation. This allows us to learn implicit shape and texture representations directly from RGB images. We experimentally show that our single-view reconstructions rival those learned with full 3D supervision. Moreover, we find that our method can be used for multi-view 3D reconstruction, directly resulting in watertight meshes.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Learning-based 3D reconstruction methods have shown impressive results. However, most methods require 3D supervision which is often hard to obtain for real-world datasets. Recently, several works have proposed differentiable rendering techniques to train reconstruction models from RGB images. Unfortunately, these approaches are currently restricted to voxel- and mesh-based representations, suffering from discretization or low resolution. In this work, we propose a differentiable rendering formulation for implicit shape and texture representations. Implicit representations have recently gained popularity as they represent shape and texture continuously. Our key insight is that depth gradients can be derived analytically using the concept of implicit differentiation. This allows us to learn implicit shape and texture representations directly from RGB images. We experimentally show that our single-view reconstructions rival those learned with full 3D supervision. Moreover, we find that our method can be used for multi-view 3D reconstruction, directly resulting in watertight meshes.", "fno": "716800d501", "keywords": [ "Gradient Methods", "Image Colour Analysis", "Image Reconstruction", "Image Representation", "Image Texture", "Learning Artificial Intelligence", "Mesh Generation", "Rendering Computer Graphics", "Shape Recognition", "Stereo Image Processing", "Texture Representations", "RGB Images", "Single View Reconstructions", "Multiview 3 D Reconstruction", "Differentiable Volumetric Rendering", "3 D Supervision", "3 D Reconstruction", "Differentiable Rendering Formulation", "Implicit 3 D Representation Learning", "Implicit Shape Representation", "Watertight Meshes", "Mesh Based Representation", "Voxel Based Representation", "Three Dimensional Displays", "Rendering Computer Graphics", "Shape", "Image Reconstruction", "Two Dimensional Displays", "Geometry", "Training" ], "authors": [ { "affiliation": "Max Planck Institute for Intelligent Systems, Tübingen; University of Tübingen", "fullName": "Michael Niemeyer", "givenName": "Michael", "surname": "Niemeyer", "__typename": "ArticleAuthorType" }, { "affiliation": "Max Planck Institute for Intelligent Systems, Tübingen; University of Tübingen; Amazon, Tübingen", "fullName": "Lars Mescheder", "givenName": "Lars", "surname": "Mescheder", "__typename": "ArticleAuthorType" }, { "affiliation": "Max Planck Institute for Intelligent Systems, Tübingen; University of Tübingen; ETAS GmbH, Bosch Group, Stuttgart", "fullName": "Michael Oechsle", "givenName": "Michael", "surname": "Oechsle", "__typename": "ArticleAuthorType" }, { "affiliation": "Max Planck Institute for Intelligent Systems, Tübingen; University of Tübingen", "fullName": "Andreas Geiger", "givenName": "Andreas", "surname": "Geiger", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-06-01T00:00:00", "pubType": "proceedings", "pages": "3501-3512", "year": "2020", "issn": null, "isbn": "978-1-7281-7168-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "716800d490", "articleId": "1m3ocuqr0wU", "__typename": "AdjacentArticleType" }, "next": { "fno": "716800d513", "articleId": "1m3nOw0NZaE", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "trans/tg/2023/06/09705143", "title": "Adaptive Joint Optimization for 3D Reconstruction With Differentiable Rendering", "doi": null, "abstractUrl": "/journal/tg/2023/06/09705143/1AIIcwNiqxq", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600e319", "title": "Recovering Fine Details for Neural Implicit Surface Reconstruction", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600e319/1KxUSVbk6He", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/01/09134794", "title": "A General Differentiable Mesh Renderer for Image-Based 3D Reasoning", "doi": null, "abstractUrl": "/journal/tp/2022/01/09134794/1lgLr0OBt9C", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800c016", "title": "DIST: Rendering Deep Implicit Signed Distance Function With Differentiable Sphere Tracing", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800c016/1m3nuEI1jJm", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a452", "title": "Learning Implicit Surface Light Fields", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a452/1qyxkR2YxGM", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a423", "title": "Semantic Implicit Neural Scene Representations With Semi-Supervised Training", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a423/1qyxoFrZU88", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900b429", "title": "Deep Implicit Templates for 3D Shape Representation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900b429/1yeLEHLhLEc", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900h115", "title": "NeuTex: Neural Texture Mapping for Volumetric Neural Rendering", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900h115/1yeLdyIKnV6", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900i928", "title": "DI-Fusion: Online Implicit 3D Reconstruction with Deep Priors", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900i928/1yeLpskgFXi", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2021/2354/0/235400a017", "title": "Neural Networks for Implicit Representations of 3D Scenes", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2021/235400a017/1zurtEgv09a", "parentPublication": { "id": "proceedings/sibgrapi/2021/2354/0", "title": "2021 34th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1qyxi3OgORy", "title": "2020 International Conference on 3D Vision (3DV)", "acronym": "3dv", "groupId": "1800494", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1qyxjJVmLQI", "doi": "10.1109/3DV50981.2020.00033", "title": "Cycle-Consistent Generative Rendering for 2D-3D Modality Translation", "normalizedTitle": "Cycle-Consistent Generative Rendering for 2D-3D Modality Translation", "abstract": "For humans, visual understanding is inherently generative: given a 3D shape, we can postulate how it would look in the world; given a 2D image, we can infer the 3D structure that likely gave rise to it. We can thus translate between the 2D visual and 3D structural modalities of a given object. In the context of computer vision, this corresponds to a learnable module that serves two purposes: (i) generate a realistic rendering of a 3D object (shape-to-image translation) and (ii) infer a realistic 3D shape from an image (image-to-shape translation). In this paper, we learn such a module while being conscious of the difficulties in obtaining large paired 2D-3D datasets. By leveraging generative domain translation methods, we are able to define a learning algorithm that requires only weak supervision, with unpaired data. The resulting model is not only able to perform 3D shape, pose, and texture inference from 2D images, but can also generate novel textured 3D shapes and renders, similar to a graphics pipeline. More specifically, our method (i) infers an explicit 3D mesh representation, (ii) utilizes example shapes to regularize inference, (iii) requires only an image mask (no keypoints or camera extrinsics), and (iv) has generative capabilities. While prior work explores subsets of these properties, their combination is novel. We demonstrate the utility of our learned representation, as well as its performance on image generation and unpaired 3D shape inference tasks.", "abstracts": [ { "abstractType": "Regular", "content": "For humans, visual understanding is inherently generative: given a 3D shape, we can postulate how it would look in the world; given a 2D image, we can infer the 3D structure that likely gave rise to it. We can thus translate between the 2D visual and 3D structural modalities of a given object. In the context of computer vision, this corresponds to a learnable module that serves two purposes: (i) generate a realistic rendering of a 3D object (shape-to-image translation) and (ii) infer a realistic 3D shape from an image (image-to-shape translation). In this paper, we learn such a module while being conscious of the difficulties in obtaining large paired 2D-3D datasets. By leveraging generative domain translation methods, we are able to define a learning algorithm that requires only weak supervision, with unpaired data. The resulting model is not only able to perform 3D shape, pose, and texture inference from 2D images, but can also generate novel textured 3D shapes and renders, similar to a graphics pipeline. More specifically, our method (i) infers an explicit 3D mesh representation, (ii) utilizes example shapes to regularize inference, (iii) requires only an image mask (no keypoints or camera extrinsics), and (iv) has generative capabilities. While prior work explores subsets of these properties, their combination is novel. We demonstrate the utility of our learned representation, as well as its performance on image generation and unpaired 3D shape inference tasks.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "For humans, visual understanding is inherently generative: given a 3D shape, we can postulate how it would look in the world; given a 2D image, we can infer the 3D structure that likely gave rise to it. We can thus translate between the 2D visual and 3D structural modalities of a given object. In the context of computer vision, this corresponds to a learnable module that serves two purposes: (i) generate a realistic rendering of a 3D object (shape-to-image translation) and (ii) infer a realistic 3D shape from an image (image-to-shape translation). In this paper, we learn such a module while being conscious of the difficulties in obtaining large paired 2D-3D datasets. By leveraging generative domain translation methods, we are able to define a learning algorithm that requires only weak supervision, with unpaired data. The resulting model is not only able to perform 3D shape, pose, and texture inference from 2D images, but can also generate novel textured 3D shapes and renders, similar to a graphics pipeline. More specifically, our method (i) infers an explicit 3D mesh representation, (ii) utilizes example shapes to regularize inference, (iii) requires only an image mask (no keypoints or camera extrinsics), and (iv) has generative capabilities. While prior work explores subsets of these properties, their combination is novel. We demonstrate the utility of our learned representation, as well as its performance on image generation and unpaired 3D shape inference tasks.", "fno": "812800a230", "keywords": [ "Computer Vision", "Data Visualisation", "Feature Extraction", "Image Reconstruction", "Image Representation", "Image Texture", "Learning Artificial Intelligence", "Mesh Generation", "Realistic Images", "Rendering Computer Graphics", "Solid Modelling", "3 D Structural Modalities", "Realistic Rendering", "Shape To Image Translation", "Realistic 3 D Shape", "Image To Shape Translation", "Generative Domain Translation Methods", "Learning Algorithm", "Texture Inference", "3 D Mesh Representation", "Image Mask", "Image Generation", "Unpaired 3 D Shape Inference Tasks", "Cycle Consistent Generative Rendering", "Visual Understanding", "Subsets", "2 D 3 D Modality Translation", "Computer Vision", "Three Dimensional Displays", "Shape", "Two Dimensional Displays", "Solid Modeling", "Image Reconstruction", "Rendering Computer Graphics", "Computational Modeling", "Generative Models", "Weak Supervision" ], "authors": [ { "affiliation": "Samsung AI Centre Toronto", "fullName": "Tristan Aumentado-Armstrong", "givenName": "Tristan", "surname": "Aumentado-Armstrong", "__typename": "ArticleAuthorType" }, { "affiliation": "Samsung AI Centre Toronto", "fullName": "Alex Levinshtein", "givenName": "Alex", "surname": "Levinshtein", "__typename": "ArticleAuthorType" }, { "affiliation": "Samsung AI Centre Toronto", "fullName": "Stavros Tsogkas", "givenName": "Stavros", "surname": "Tsogkas", "__typename": "ArticleAuthorType" }, { "affiliation": "Samsung AI Centre Toronto", "fullName": "Konstantinos G. Derpanis", "givenName": "Konstantinos G.", "surname": "Derpanis", "__typename": "ArticleAuthorType" }, { "affiliation": "Samsung AI Centre Toronto", "fullName": "Allan D. Jepson", "givenName": "Allan D.", "surname": "Jepson", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-11-01T00:00:00", "pubType": "proceedings", "pages": "230-240", "year": "2020", "issn": null, "isbn": "978-1-7281-8128-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "812800a220", "articleId": "1qyxlQVwTKM", "__typename": "AdjacentArticleType" }, "next": { "fno": "812800a241", "articleId": "1qyxlfp4xG0", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dv/2017/2610/0/261001a402", "title": "3D Shape Induction from 2D Views of Multiple Objects", "doi": null, "abstractUrl": "/proceedings-article/3dv/2017/261001a402/12OmNBLdKOR", "parentPublication": { "id": "proceedings/3dv/2017/2610/0", "title": "2017 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2017/1032/0/1032a853", "title": "A Generative Model of People in Clothing", "doi": null, "abstractUrl": "/proceedings-article/iccv/2017/1032a853/12OmNCwlakU", "parentPublication": { "id": "proceedings/iccv/2017/1032/0", "title": "2017 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2017/0457/0/0457a388", "title": "Deep Supervision with Shape Concepts for Occlusion-Aware 3D Object Parsing", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457a388/12OmNx5piPM", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2018/09/08031057", "title": "Viewpoint-Consistent 3D Face Alignment", "doi": null, "abstractUrl": "/journal/tp/2018/09/08031057/13rRUxcbnDR", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600s8429", "title": "Multi-View Consistent Generative Adversarial Networks for 3D-aware Image Synthesis", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600s8429/1H1hRTUHoLS", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300h790", "title": "View Independent Generative Adversarial Network for Novel View Synthesis", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300h790/1hVlLtjahSE", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2020/9360/0/09150810", "title": "Estimation of Orientation and Camera Parameters from Cryo-Electron Microscopy Images with Variational Autoencoders and Generative Adversarial Networks", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2020/09150810/1lPHcyiKALu", "parentPublication": { "id": "proceedings/cvprw/2020/9360/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800f870", "title": "Towards Unsupervised Learning of Generative Models for 3D Controllable Image Synthesis", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800f870/1m3nWAV8Arm", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a868", "title": "GIF: Generative Interpretable Faces", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a868/1qyxnIhctWg", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800b039", "title": "Self-Supervised 2D Image to 3D Shape Translation with Disentangled Representations", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800b039/1qyxokYBKrC", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyoiYVr", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNB0FxiX", "doi": "10.1109/CVPR.2017.385", "title": "Learning Barycentric Representations of 3D Shapes for Sketch-Based 3D Shape Retrieval", "normalizedTitle": "Learning Barycentric Representations of 3D Shapes for Sketch-Based 3D Shape Retrieval", "abstract": "Retrieving 3D shapes with sketches is a challenging problem since 2D sketches and 3D shapes are from two heterogeneous domains, which results in large discrepancy between them. In this paper, we propose to learn barycenters of 2D projections of 3D shapes for sketch-based 3D shape retrieval. Specifically, we first use two deep convolutional neural networks (CNNs) to extract deep features of sketches and 2D projections of 3D shapes. For 3D shapes, we then compute the Wasserstein barycenters of deep features of multiple projections to form a barycentric representation. Finally, by constructing a metric network, a discriminative loss is formulated on the Wasserstein barycenters of 3D shapes and sketches in the deep feature space to learn discriminative and compact 3D shape and sketch features for retrieval. The proposed method is evaluated on the SHREC13 and SHREC14 sketch track benchmark datasets. Compared to the state-of-the-art methods, our proposed method can significantly improve the retrieval performance.", "abstracts": [ { "abstractType": "Regular", "content": "Retrieving 3D shapes with sketches is a challenging problem since 2D sketches and 3D shapes are from two heterogeneous domains, which results in large discrepancy between them. In this paper, we propose to learn barycenters of 2D projections of 3D shapes for sketch-based 3D shape retrieval. Specifically, we first use two deep convolutional neural networks (CNNs) to extract deep features of sketches and 2D projections of 3D shapes. For 3D shapes, we then compute the Wasserstein barycenters of deep features of multiple projections to form a barycentric representation. Finally, by constructing a metric network, a discriminative loss is formulated on the Wasserstein barycenters of 3D shapes and sketches in the deep feature space to learn discriminative and compact 3D shape and sketch features for retrieval. The proposed method is evaluated on the SHREC13 and SHREC14 sketch track benchmark datasets. Compared to the state-of-the-art methods, our proposed method can significantly improve the retrieval performance.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Retrieving 3D shapes with sketches is a challenging problem since 2D sketches and 3D shapes are from two heterogeneous domains, which results in large discrepancy between them. In this paper, we propose to learn barycenters of 2D projections of 3D shapes for sketch-based 3D shape retrieval. Specifically, we first use two deep convolutional neural networks (CNNs) to extract deep features of sketches and 2D projections of 3D shapes. For 3D shapes, we then compute the Wasserstein barycenters of deep features of multiple projections to form a barycentric representation. Finally, by constructing a metric network, a discriminative loss is formulated on the Wasserstein barycenters of 3D shapes and sketches in the deep feature space to learn discriminative and compact 3D shape and sketch features for retrieval. The proposed method is evaluated on the SHREC13 and SHREC14 sketch track benchmark datasets. Compared to the state-of-the-art methods, our proposed method can significantly improve the retrieval performance.", "fno": "0457d615", "keywords": [ "Feature Extraction", "Image Representation", "Image Retrieval", "Learning Artificial Intelligence", "Neural Nets", "Solid Modelling", "Wasserstein Barycenters", "Deep Feature Space", "Deep Convolutional Neural Networks", "Barycentric Representation Learning", "Sketch Based 3 D Shape Retrieval", "2 D Projections", "CNN", "Deep Feature Extraction", "Metric Network", "Discriminative Loss", "SHREC 13 Sketch Track Benchmark Dataset", "SHREC 14 Sketch Track Benchmark Dataset", "Shape", "Three Dimensional Displays", "Two Dimensional Displays", "Feature Extraction", "Neural Networks", "Probability Distribution", "Visualization" ], "authors": [ { "affiliation": null, "fullName": "Jin Xie", "givenName": "Jin", "surname": "Xie", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Guoxian Dai", "givenName": "Guoxian", "surname": "Dai", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Fan Zhu", "givenName": "Fan", "surname": "Zhu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yi Fang", "givenName": "Yi", "surname": "Fang", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-07-01T00:00:00", "pubType": "proceedings", "pages": "3615-3623", "year": "2017", "issn": "1063-6919", "isbn": "978-1-5386-0457-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0457d605", "articleId": "12OmNywfKFh", "__typename": "AdjacentArticleType" }, "next": { "fno": "0457d624", "articleId": "12OmNx38vRo", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icme/2017/6067/0/08019464", "title": "Multi-view pairwise relationship learning for sketch based 3D shape retrieval", "doi": null, "abstractUrl": "/proceedings-article/icme/2017/08019464/12OmNy6Zs2q", "parentPublication": { "id": "proceedings/icme/2017/6067/0", "title": "2017 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446609", "title": "Model Retrieval by 3D Sketching in Immersive Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446609/13bd1tl2omk", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2017/06/mcg2017060088", "title": "Sketch-Based Articulated 3D Shape Retrieval", "doi": null, "abstractUrl": "/magazine/cg/2017/06/mcg2017060088/13rRUwfqpG7", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2018/9497/0/949700a311", "title": "Sketch-Based Shape Retrieval via Multi-view Attention and Generalized Similarity", "doi": null, "abstractUrl": "/proceedings-article/icdh/2018/949700a311/17D45VObpQZ", "parentPublication": { "id": "proceedings/icdh/2018/9497/0", "title": "2018 7th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2022/5670/0/567000a383", "title": "Structure-Aware 3D VR Sketch to 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/3dv/2022/567000a383/1KYsqgmUniE", "parentPublication": { "id": "proceedings/3dv/2022/5670/0", "title": "2022 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08964443", "title": "DeepSketchHair: Deep Sketch-Based 3D Hair Modeling", "doi": null, "abstractUrl": "/journal/tg/2021/07/08964443/1gLZSnCp3Ko", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2016/4847/0/07900083", "title": "3D sketch-based 3D model retrieval with convolutional neural network", "doi": null, "abstractUrl": "/proceedings-article/icpr/2016/07900083/1gysq8EnfHi", "parentPublication": { "id": "proceedings/icpr/2016/4847/0", "title": "2016 23rd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2020/1331/0/09102925", "title": "Cross-Modal Guidance Network For Sketch-Based 3d Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/icme/2020/09102925/1kwqTrDSXF6", "parentPublication": { "id": "proceedings/icme/2020/1331/0", "title": "2020 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a081", "title": "Towards 3D VR-Sketch to 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a081/1qyxlDtR0Ji", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2020/9234/0/923400a184", "title": "Deep 3D Shape Reconstruction from Single-View Sketch Image", "doi": null, "abstractUrl": "/proceedings-article/icdh/2020/923400a184/1uGY2GTiIda", "parentPublication": { "id": "proceedings/icdh/2020/9234/0", "title": "2020 8th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "17D45VtKisL", "title": "2018 7th International Conference on Digital Home (ICDH)", "acronym": "icdh", "groupId": "1802037", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "17D45VObpQZ", "doi": "10.1109/ICDH.2018.00061", "title": "Sketch-Based Shape Retrieval via Multi-view Attention and Generalized Similarity", "normalizedTitle": "Sketch-Based Shape Retrieval via Multi-view Attention and Generalized Similarity", "abstract": "Sketch-based shape retrieval has received increasing attention in computer vision and computer graphics. It suffers from the challenge gap between 2D sketches and 3D shapes. In this paper, we propose a generalized similarity matching framework based on a multi-view attention network (MVAN), which can retrieve 3D shape that is most similar to the query sketch. In proposed approach, firstly we compute 2D projections of 3D shapes from multiple viewpoints and utilize a convolutional neural network to extract low level feature maps of these 2D projections. Secondly a multi-view attention network is designed to fuse the feature maps and forms a more accurate 3D shape representation. Meanwhile we use a CNN to extract the feature of sketches. Thirdly the similarity between sketches and 3D shapes is estimated via a generalized similarity model, which fuses some traditional similarity model into a generalized form and optimizes its parameters using a data-driven method. Finally we combine the MVAN and generalized similarity model into a unified network and train the model in an end-to-end manner. The experimental results on SHREC'13 and SHREC'14 sketch track benchmark datasets demonstrate that the proposed method can outperform state-of-the-art methods.", "abstracts": [ { "abstractType": "Regular", "content": "Sketch-based shape retrieval has received increasing attention in computer vision and computer graphics. It suffers from the challenge gap between 2D sketches and 3D shapes. In this paper, we propose a generalized similarity matching framework based on a multi-view attention network (MVAN), which can retrieve 3D shape that is most similar to the query sketch. In proposed approach, firstly we compute 2D projections of 3D shapes from multiple viewpoints and utilize a convolutional neural network to extract low level feature maps of these 2D projections. Secondly a multi-view attention network is designed to fuse the feature maps and forms a more accurate 3D shape representation. Meanwhile we use a CNN to extract the feature of sketches. Thirdly the similarity between sketches and 3D shapes is estimated via a generalized similarity model, which fuses some traditional similarity model into a generalized form and optimizes its parameters using a data-driven method. Finally we combine the MVAN and generalized similarity model into a unified network and train the model in an end-to-end manner. The experimental results on SHREC'13 and SHREC'14 sketch track benchmark datasets demonstrate that the proposed method can outperform state-of-the-art methods.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Sketch-based shape retrieval has received increasing attention in computer vision and computer graphics. It suffers from the challenge gap between 2D sketches and 3D shapes. In this paper, we propose a generalized similarity matching framework based on a multi-view attention network (MVAN), which can retrieve 3D shape that is most similar to the query sketch. In proposed approach, firstly we compute 2D projections of 3D shapes from multiple viewpoints and utilize a convolutional neural network to extract low level feature maps of these 2D projections. Secondly a multi-view attention network is designed to fuse the feature maps and forms a more accurate 3D shape representation. Meanwhile we use a CNN to extract the feature of sketches. Thirdly the similarity between sketches and 3D shapes is estimated via a generalized similarity model, which fuses some traditional similarity model into a generalized form and optimizes its parameters using a data-driven method. Finally we combine the MVAN and generalized similarity model into a unified network and train the model in an end-to-end manner. The experimental results on SHREC'13 and SHREC'14 sketch track benchmark datasets demonstrate that the proposed method can outperform state-of-the-art methods.", "fno": "949700a311", "keywords": [ "Computer Vision", "Convolutional Neural Nets", "Feature Extraction", "Image Matching", "Image Representation", "Image Retrieval", "Shape Recognition", "Solid Modelling", "Sketch Based Shape Retrieval", "Computer Vision", "Computer Graphics", "Generalized Similarity Matching Framework", "Multiview Attention Network", "Query Sketch", "Convolutional Neural Network", "Low Level Feature Maps", "Accurate 3 D Shape Representation", "Generalized Similarity Model", "Traditional Similarity Model", "Generalized Form", "SHREC 14 Sketch Track Benchmark Datasets", "MVAN", "CNN", "2 D Sketches", "Shape", "Three Dimensional Displays", "Feature Extraction", "Fuses", "Solid Modeling", "Shape Measurement", "Two Dimensional Displays", "Sketch Shape Attention Generalized Similarity" ], "authors": [ { "affiliation": null, "fullName": "Yongzhe Xu", "givenName": "Yongzhe", "surname": "Xu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jiangchuan Hu", "givenName": "Jiangchuan", "surname": "Hu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Kun Zeng", "givenName": "Kun", "surname": "Zeng", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yongyi Gong", "givenName": "Yongyi", "surname": "Gong", "__typename": "ArticleAuthorType" } ], "idPrefix": "icdh", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-11-01T00:00:00", "pubType": "proceedings", "pages": "311-317", "year": "2018", "issn": null, "isbn": "978-1-5386-9497-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "949700a305", "articleId": "17D45WWzW2N", "__typename": "AdjacentArticleType" }, "next": { "fno": "949700a318", "articleId": "17D45XzbnL1", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icme/2017/6067/0/08019464", "title": "Multi-view pairwise relationship learning for sketch based 3D shape retrieval", "doi": null, "abstractUrl": "/proceedings-article/icme/2017/08019464/12OmNy6Zs2q", "parentPublication": { "id": "proceedings/icme/2017/6067/0", "title": "2017 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2014/4284/0/4284a277", "title": "Sketch-Based Shape Retrieval Using Orientation Histogram with Gabor Filters", "doi": null, "abstractUrl": "/proceedings-article/icdh/2014/4284a277/12OmNzlUKfK", "parentPublication": { "id": "proceedings/icdh/2014/4284/0", "title": "2014 5th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2017/06/mcg2017060088", "title": "Sketch-Based Articulated 3D Shape Retrieval", "doi": null, "abstractUrl": "/magazine/cg/2017/06/mcg2017060088/13rRUwfqpG7", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2022/5670/0/567000a383", "title": "Structure-Aware 3D VR Sketch to 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/3dv/2022/567000a383/1KYsqgmUniE", "parentPublication": { "id": "proceedings/3dv/2022/5670/0", "title": "2022 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2022/5670/0/567000a022", "title": "Garment Ideation: Iterative View-Aware Sketch-Based Garment Modeling", "doi": null, "abstractUrl": "/proceedings-article/3dv/2022/567000a022/1KYsti3axvq", "parentPublication": { "id": "proceedings/3dv/2022/5670/0", "title": "2022 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/08/09007505", "title": "Sketch Augmentation-Driven Shape Retrieval Learning Framework Based on Convolutional Neural Networks", "doi": null, "abstractUrl": "/journal/tg/2021/08/09007505/1hJKlMJzueI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2020/1331/0/09102925", "title": "Cross-Modal Guidance Network For Sketch-Based 3d Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/icme/2020/09102925/1kwqTrDSXF6", "parentPublication": { "id": "proceedings/icme/2020/1331/0", "title": "2020 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a081", "title": "Towards 3D VR-Sketch to 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a081/1qyxlDtR0Ji", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2020/9234/0/923400a184", "title": "Deep 3D Shape Reconstruction from Single-View Sketch Image", "doi": null, "abstractUrl": "/proceedings-article/icdh/2020/923400a184/1uGY2GTiIda", "parentPublication": { "id": "proceedings/icdh/2020/9234/0", "title": "2020 8th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icicta/2020/8666/0/866600a223", "title": "Sketch-based 3D Shape Retrieval with Multi-Silhouette View Based on Convolutional Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icicta/2020/866600a223/1wRIvGNgH9m", "parentPublication": { "id": "proceedings/icicta/2020/8666/0", "title": "2020 13th International Conference on Intelligent Computation Technology and Automation (ICICTA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1a3x4M4IIJa", "title": "2018 International Conference on Virtual Reality and Visualization (ICVRV)", "acronym": "icvrv", "groupId": "1800579", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "1a3x7jLsFPi", "doi": "10.1109/ICVRV.2018.00010", "title": "Data-Driven Hair Modeling from a Single Image", "normalizedTitle": "Data-Driven Hair Modeling from a Single Image", "abstract": "Hair is one of the most distinctive part of a human body, which is essential for the digitization of compelling virtual avatars. We present a data-driven approach for generating complete and complex 3D hairstyles from a single-view portrait. We construct a hairstyle database which contains more than 2000 hair models. Given a target hairstyle portrait image, we first draw a few strokes as guidance. We then search multiple best matching examples from the database based on our enhanced matching algorithm. Finally, we combine them consistently into a single hairstyle. The generated hairstyles are visually comparable to original portrait images. The reconstructed 3D hair models can be used for many applications, such as hair editing, and dynamic hair simulation.", "abstracts": [ { "abstractType": "Regular", "content": "Hair is one of the most distinctive part of a human body, which is essential for the digitization of compelling virtual avatars. We present a data-driven approach for generating complete and complex 3D hairstyles from a single-view portrait. We construct a hairstyle database which contains more than 2000 hair models. Given a target hairstyle portrait image, we first draw a few strokes as guidance. We then search multiple best matching examples from the database based on our enhanced matching algorithm. Finally, we combine them consistently into a single hairstyle. The generated hairstyles are visually comparable to original portrait images. The reconstructed 3D hair models can be used for many applications, such as hair editing, and dynamic hair simulation.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Hair is one of the most distinctive part of a human body, which is essential for the digitization of compelling virtual avatars. We present a data-driven approach for generating complete and complex 3D hairstyles from a single-view portrait. We construct a hairstyle database which contains more than 2000 hair models. Given a target hairstyle portrait image, we first draw a few strokes as guidance. We then search multiple best matching examples from the database based on our enhanced matching algorithm. Finally, we combine them consistently into a single hairstyle. The generated hairstyles are visually comparable to original portrait images. The reconstructed 3D hair models can be used for many applications, such as hair editing, and dynamic hair simulation.", "fno": "849700a008", "keywords": [ "Avatars", "Computer Animation", "Realistic Images", "Solid Modelling", "Original Portrait Images", "Hair Models", "Generated Hairstyles", "Single Hairstyle", "Enhanced Matching Algorithm", "Multiple Best Matching Examples", "Target Hairstyle Portrait Image", "Hairstyle Database", "Single View Portrait", "Complex 3 D Hairstyles", "Data Driven Approach", "Compelling Virtual Avatars", "Digitization", "Human Body", "Single Image", "Data Driven Hair Modeling", "Dynamic Hair Simulation", "Hair Editing", "Hair", "Solid Modeling", "Three Dimensional Displays", "Databases", "Computational Modeling", "Two Dimensional Displays", "Head", "Hair Modeling", "Data Driven", "Hairstyle Database", "Orientation Field" ], "authors": [ { "affiliation": null, "fullName": "Jiqiang Wu", "givenName": "Jiqiang", "surname": "Wu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yongtang Bao", "givenName": "Yongtang", "surname": "Bao", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yue Qi", "givenName": "Yue", "surname": "Qi", "__typename": "ArticleAuthorType" } ], "idPrefix": "icvrv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-10-01T00:00:00", "pubType": "proceedings", "pages": "8-14", "year": "2018", "issn": "2375-141X", "isbn": "978-1-5386-8497-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "849700a001", "articleId": "1a3x6hGWsso", "__typename": "AdjacentArticleType" }, "next": { "fno": "849700a015", "articleId": "1a3x6Frzph6", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/casa/2003/1934/0/19340041", "title": "Modeling Hair Using Level-of-Detail Representations", "doi": null, "abstractUrl": "/proceedings-article/casa/2003/19340041/12OmNA1DMn9", "parentPublication": { "id": "proceedings/casa/2003/1934/0", "title": "Computer Animation and Social Agents, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/simultech/2014/060/0/07095029", "title": "2D hair strands generation based on template matching", "doi": null, "abstractUrl": "/proceedings-article/simultech/2014/07095029/12OmNx5GU7n", "parentPublication": { "id": "proceedings/simultech/2014/060/0", "title": "2014 International Conference on Simulation and Modeling Methodologies, Technologies and Applications (SIMULTECH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2017/2089/0/2089a214", "title": "NPR Hair Modeling with Parametric Clumps", "doi": null, "abstractUrl": "/proceedings-article/cw/2017/2089a214/12OmNxuo0jO", "parentPublication": { "id": "proceedings/cw/2017/2089/0", "title": "2017 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/6.946E222", "title": "HairMapper: Removing Hair from Portraits Using GANs", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/6.946E222/1H1kQpBVCZW", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08964443", "title": "DeepSketchHair: Deep Sketch-Based 3D Hair Modeling", "doi": null, "abstractUrl": "/journal/tg/2021/07/08964443/1gLZSnCp3Ko", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nicoint/2020/8771/0/09122356", "title": "Viewpoint Selection for Sketch-based Hairstyle Modeling", "doi": null, "abstractUrl": "/proceedings-article/nicoint/2020/09122356/1kRSfSP7OpO", "parentPublication": { "id": "proceedings/nicoint/2020/8771/0", "title": "2020 Nicograph International (NicoInt)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800h444", "title": "Intuitive, Interactive Beard and Hair Synthesis With Generative Models", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800h444/1m3ol6XnJ3q", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2019/4752/0/09212824", "title": "Automatic Hair Modeling from One Image", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2019/09212824/1nHRUrDMgE0", "parentPublication": { "id": "proceedings/icvrv/2019/4752/0", "title": "2019 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/04/09220808", "title": "Real-Time Hair Simulation With Neural Interpolation", "doi": null, "abstractUrl": "/journal/tg/2022/04/09220808/1nRLElyFvfG", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900b984", "title": "LOHO: Latent Optimization of Hairstyles via Orthogonalization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900b984/1yeIuaT2Ife", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "19wADSkufD2", "title": "2016 23rd International Conference on Pattern Recognition (ICPR)", "acronym": "icpr", "groupId": "1000545", "volume": "0", "displayVolume": "0", "year": "2016", "__typename": "ProceedingType" }, "article": { "id": "1gysq8EnfHi", "doi": "10.1109/ICPR.2016.7900083", "title": "3D sketch-based 3D model retrieval with convolutional neural network", "normalizedTitle": "3D sketch-based 3D model retrieval with convolutional neural network", "abstract": "3D sketch-based 3D model retrieval is to retrieve similar 3D models using users' hand-drawn 3D sketches as input. Compared with traditional 2D sketch-based retrieval, 3D sketch-based 3D model retrieval is a brand new and challenging research topic. In this paper, we employ advanced deep learning method and propose a novel 3D sketch based 3D model retrieval system. Our system has been comprehensively tested on two benchmark datasets and compared with other existing 3D model retrieval algorithms. The experimental results reveal our approach outperforms other competing state-of-the-arts and demonstrate promising potential of our approach on 3D sketch based applications.", "abstracts": [ { "abstractType": "Regular", "content": "3D sketch-based 3D model retrieval is to retrieve similar 3D models using users' hand-drawn 3D sketches as input. Compared with traditional 2D sketch-based retrieval, 3D sketch-based 3D model retrieval is a brand new and challenging research topic. In this paper, we employ advanced deep learning method and propose a novel 3D sketch based 3D model retrieval system. Our system has been comprehensively tested on two benchmark datasets and compared with other existing 3D model retrieval algorithms. The experimental results reveal our approach outperforms other competing state-of-the-arts and demonstrate promising potential of our approach on 3D sketch based applications.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "3D sketch-based 3D model retrieval is to retrieve similar 3D models using users' hand-drawn 3D sketches as input. Compared with traditional 2D sketch-based retrieval, 3D sketch-based 3D model retrieval is a brand new and challenging research topic. In this paper, we employ advanced deep learning method and propose a novel 3D sketch based 3D model retrieval system. Our system has been comprehensively tested on two benchmark datasets and compared with other existing 3D model retrieval algorithms. The experimental results reveal our approach outperforms other competing state-of-the-arts and demonstrate promising potential of our approach on 3D sketch based applications.", "fno": "07900083", "keywords": [ "Convolution", "Information Retrieval", "Learning Artificial Intelligence", "Neural Nets", "Solid Modelling", "Convolutional Neural Network", "3 D Sketch Based 3 D Model Retrieval", "Hand Drawn 3 D Sketches", "Deep Learning", "Three Dimensional Displays", "Solid Modeling", "Two Dimensional Displays", "Computational Modeling", "Benchmark Testing", "Shape", "Training" ], "authors": [ { "affiliation": "Dept. of Comput. Sci., Texas State Univ., San Marcos, TX, USA", "fullName": "Yuxiang Ye", "givenName": "Yuxiang", "surname": "Ye", "__typename": "ArticleAuthorType" }, { "affiliation": "1 School of Computing, University of Southern Mississippi, Long Beach, 39560, USA", "fullName": "Bo Li", "givenName": "Bo", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "Department of Computer Science, Texas State University, San Marcos, 78666, USA", "fullName": "Yijuan Lu", "givenName": "Yijuan", "surname": "Lu", "__typename": "ArticleAuthorType" } ], "idPrefix": "icpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2016-12-01T00:00:00", "pubType": "proceedings", "pages": "2936-2941", "year": "2016", "issn": null, "isbn": "978-1-5090-4847-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07900064", "articleId": "1gyspWXpJuw", "__typename": "AdjacentArticleType" }, "next": { "fno": "07900084", "articleId": "1AUoXldO62c", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cw/2013/2246/0/2246a282", "title": "View-Clustering and Manifold Learning for Sketch-Based 3D Model Retrieval", "doi": null, "abstractUrl": "/proceedings-article/cw/2013/2246a282/12OmNAYXWLN", "parentPublication": { "id": "proceedings/cw/2013/2246/0", "title": "2013 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2017/0457/0/0457d615", "title": "Learning Barycentric Representations of 3D Shapes for Sketch-Based 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457d615/12OmNB0FxiX", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209e570", "title": "Sketch-Based 3D Model Retrieval via Multi-feature Fusion", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209e570/12OmNBOCWrV", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2016/4400/0/4400a261", "title": "3D Model Retrieval Based on Hand Drawn Sketches Using LDA Model", "doi": null, "abstractUrl": "/proceedings-article/icdh/2016/4400a261/12OmNqyUUvj", "parentPublication": { "id": "proceedings/icdh/2016/4400/0", "title": "2016 6th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2013/2246/0/2246a274", "title": "Ranking on Cross-Domain Manifold for Sketch-Based 3D Model Retrieval", "doi": null, "abstractUrl": "/proceedings-article/cw/2013/2246a274/12OmNvDZEUr", "parentPublication": { "id": "proceedings/cw/2013/2246/0", "title": "2013 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2016/4571/0/4571a173", "title": "Similarity Retrieval of 3D Models with Query by Clay Sketch", "doi": null, "abstractUrl": "/proceedings-article/ism/2016/4571a173/12OmNvk7JML", "parentPublication": { "id": "proceedings/ism/2016/4571/0", "title": "2016 IEEE International Symposium on Multimedia (ISM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2017/06/mcg2017060088", "title": "Sketch-Based Articulated 3D Shape Retrieval", "doi": null, "abstractUrl": "/magazine/cg/2017/06/mcg2017060088/13rRUwfqpG7", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/08/09007505", "title": "Sketch Augmentation-Driven Shape Retrieval Learning Framework Based on Convolutional Neural Networks", "doi": null, "abstractUrl": "/journal/tg/2021/08/09007505/1hJKlMJzueI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a081", "title": "Towards 3D VR-Sketch to 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a081/1qyxlDtR0Ji", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icicta/2020/8666/0/866600a223", "title": "Sketch-based 3D Shape Retrieval with Multi-Silhouette View Based on Convolutional Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icicta/2020/866600a223/1wRIvGNgH9m", "parentPublication": { "id": "proceedings/icicta/2020/8666/0", "title": "2020 13th International Conference on Intelligent Computation Technology and Automation (ICICTA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1kRSe09ZlTO", "title": "2020 Nicograph International (NicoInt)", "acronym": "nicoint", "groupId": "1814784", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1kRSfSP7OpO", "doi": "10.1109/NicoInt50878.2020.00024", "title": "Viewpoint Selection for Sketch-based Hairstyle Modeling", "normalizedTitle": "Viewpoint Selection for Sketch-based Hairstyle Modeling", "abstract": "Hairstyle is an essential factor for representing individuals. It is in demand for ordinary persons who are not experts of professional software on three-dimensional computer graphics (3DCG) to represent hairstyles with flexible hairstyle modeling using 3DCG. We propose a method to realize sketch-based hairstyle modeling efficiently by supporting a viewpoint recommendation technique so that such users can easily perform hairstyle modeling. This method automatically finds appropriate viewpoints by evaluating the shapes of sketches on the projected 2D space. We aim to make this task of sketch-based hairstyle modeling efficient and straightforward because the technique avoids to manually control the viewpoint and prompts users to perform hair modeling intuitively. This paper introduces examples of viewpoint recommendation for two different input sketches.", "abstracts": [ { "abstractType": "Regular", "content": "Hairstyle is an essential factor for representing individuals. It is in demand for ordinary persons who are not experts of professional software on three-dimensional computer graphics (3DCG) to represent hairstyles with flexible hairstyle modeling using 3DCG. We propose a method to realize sketch-based hairstyle modeling efficiently by supporting a viewpoint recommendation technique so that such users can easily perform hairstyle modeling. This method automatically finds appropriate viewpoints by evaluating the shapes of sketches on the projected 2D space. We aim to make this task of sketch-based hairstyle modeling efficient and straightforward because the technique avoids to manually control the viewpoint and prompts users to perform hair modeling intuitively. This paper introduces examples of viewpoint recommendation for two different input sketches.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Hairstyle is an essential factor for representing individuals. It is in demand for ordinary persons who are not experts of professional software on three-dimensional computer graphics (3DCG) to represent hairstyles with flexible hairstyle modeling using 3DCG. We propose a method to realize sketch-based hairstyle modeling efficiently by supporting a viewpoint recommendation technique so that such users can easily perform hairstyle modeling. This method automatically finds appropriate viewpoints by evaluating the shapes of sketches on the projected 2D space. We aim to make this task of sketch-based hairstyle modeling efficient and straightforward because the technique avoids to manually control the viewpoint and prompts users to perform hair modeling intuitively. This paper introduces examples of viewpoint recommendation for two different input sketches.", "fno": "09122356", "keywords": [ "Computer Animation", "Curve Fitting", "Rendering Computer Graphics", "Solid Modelling", "Flexible Hairstyle Modeling", "3 DCG", "Viewpoint Recommendation Technique", "Viewpoint Selection", "Three Dimensional Computer Graphics", "Input Sketches", "Hair", "Solid Modeling", "Shape", "Computational Modeling", "Computer Graphics", "Aerospace Electronics", "Software", "Component", "Formatting", "Style", "Styling" ], "authors": [ { "affiliation": "Ochanomizu University,Tokyo,Japan", "fullName": "Moeko Ishii", "givenName": "Moeko", "surname": "Ishii", "__typename": "ArticleAuthorType" }, { "affiliation": "Ochanomizu University,Tokyo,Japan", "fullName": "Takayuki Itoh", "givenName": "Takayuki", "surname": "Itoh", "__typename": "ArticleAuthorType" } ], "idPrefix": "nicoint", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-06-01T00:00:00", "pubType": "proceedings", "pages": "82-85", "year": "2020", "issn": null, "isbn": "978-1-7281-8771-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09122333", "articleId": "1kRSff500fK", "__typename": "AdjacentArticleType" }, "next": { "fno": "09122336", "articleId": "1kRSelKDDk4", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cgiv/2014/5720/0/5720a027", "title": "A Survey of Sketch Based Modeling Systems", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2014/5720a027/12OmNwLOYTE", "parentPublication": { "id": "proceedings/cgiv/2014/5720/0", "title": "2014 11th International Conference on Computer Graphics, Imaging and Visualization (CGIV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgi/2004/2171/0/21710608", "title": "Sketch Interface Based Expressive Hairstyle Modelling and Rendering", "doi": null, "abstractUrl": "/proceedings-article/cgi/2004/21710608/12OmNweBUN2", "parentPublication": { "id": "proceedings/cgi/2004/2171/0", "title": "Proceedings. Computer Graphics International", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2017/2089/0/2089a214", "title": "NPR Hair Modeling with Parametric Clumps", "doi": null, "abstractUrl": "/proceedings-article/cw/2017/2089a214/12OmNxuo0jO", "parentPublication": { "id": "proceedings/cw/2017/2089/0", "title": "2017 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmi/2002/1834/0/18340535", "title": "An Improved Algorithm for Hairstyle Dynamics", "doi": null, "abstractUrl": "/proceedings-article/icmi/2002/18340535/12OmNxwWoUQ", "parentPublication": { "id": "proceedings/icmi/2002/1834/0", "title": "Proceedings Fourth IEEE International Conference on Multimodal Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgiv/2006/2606/0/26060365", "title": "Hairstyle Construction from Raw Surface Data", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2006/26060365/12OmNzgwmL5", "parentPublication": { "id": "proceedings/cgiv/2006/2606/0", "title": "International Conference on Computer Graphics, Imaging and Visualisation (CGIV'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fg/2021/3176/0/09667038", "title": "Hairstyle Transfer between Face Images", "doi": null, "abstractUrl": "/proceedings-article/fg/2021/09667038/1A6BEOBJyq4", "parentPublication": { "id": "proceedings/fg/2021/3176/0", "title": "2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2018/8497/0/849700a008", "title": "Data-Driven Hair Modeling from a Single Image", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2018/849700a008/1a3x7jLsFPi", "parentPublication": { "id": "proceedings/icvrv/2018/8497/0", "title": "2018 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08964443", "title": "DeepSketchHair: Deep Sketch-Based 3D Hair Modeling", "doi": null, "abstractUrl": "/journal/tg/2021/07/08964443/1gLZSnCp3Ko", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2019/4752/0/09212824", "title": "Automatic Hair Modeling from One Image", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2019/09212824/1nHRUrDMgE0", "parentPublication": { "id": "proceedings/icvrv/2019/4752/0", "title": "2019 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a543", "title": "Deep Sketch-Based Modeling: Tips and Tricks", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a543/1qyxmtKBkLS", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1kwqNHC4Fy0", "title": "2020 IEEE International Conference on Multimedia and Expo (ICME)", "acronym": "icme", "groupId": "1000477", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1kwqTrDSXF6", "doi": "10.1109/ICME46284.2020.9102925", "title": "Cross-Modal Guidance Network For Sketch-Based 3d Shape Retrieval", "normalizedTitle": "Cross-Modal Guidance Network For Sketch-Based 3d Shape Retrieval", "abstract": "The main challenge of sketch-based 3D shape retrieval is the large cross-modal differences between 2D sketches and 3D shapes. Most recent works employed two heterogeneous networks and a shared loss to directly map the features from different modalities to a common feature space, which failed to reduce the cross-modal differences effectively. In this paper, we propose a novel method that adopts a teacher-student strategy to learn an aligned cross-modal feature space indirectly. Specifically, our method first employs a classification network to learn the discriminative features of 3D shapes. Then, the pre-learned features are considered as a teacher to guide the feature learning of 2D sketches. In order to align the cross-modal features, 2D sketch features are transferred to the pre-learned 3D feature space. Our experiments on two benchmark datasets demonstrate that our method obtains superior retrieval performance than the state-of-the-art approaches.", "abstracts": [ { "abstractType": "Regular", "content": "The main challenge of sketch-based 3D shape retrieval is the large cross-modal differences between 2D sketches and 3D shapes. Most recent works employed two heterogeneous networks and a shared loss to directly map the features from different modalities to a common feature space, which failed to reduce the cross-modal differences effectively. In this paper, we propose a novel method that adopts a teacher-student strategy to learn an aligned cross-modal feature space indirectly. Specifically, our method first employs a classification network to learn the discriminative features of 3D shapes. Then, the pre-learned features are considered as a teacher to guide the feature learning of 2D sketches. In order to align the cross-modal features, 2D sketch features are transferred to the pre-learned 3D feature space. Our experiments on two benchmark datasets demonstrate that our method obtains superior retrieval performance than the state-of-the-art approaches.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The main challenge of sketch-based 3D shape retrieval is the large cross-modal differences between 2D sketches and 3D shapes. Most recent works employed two heterogeneous networks and a shared loss to directly map the features from different modalities to a common feature space, which failed to reduce the cross-modal differences effectively. In this paper, we propose a novel method that adopts a teacher-student strategy to learn an aligned cross-modal feature space indirectly. Specifically, our method first employs a classification network to learn the discriminative features of 3D shapes. Then, the pre-learned features are considered as a teacher to guide the feature learning of 2D sketches. In order to align the cross-modal features, 2D sketch features are transferred to the pre-learned 3D feature space. Our experiments on two benchmark datasets demonstrate that our method obtains superior retrieval performance than the state-of-the-art approaches.", "fno": "09102925", "keywords": [ "Feature Extraction", "Image Retrieval", "Learning Artificial Intelligence", "Solid Modelling", "Pre Learned 3 D Feature Space", "2 D Sketch Features", "Cross Modal Features", "Feature Learning", "Pre Learned Features", "Discriminative Features", "Aligned Cross Modal Feature Space", "Common Feature Space", "Heterogeneous Networks", "Cross Modal Differences", "Sketch Based 3 D Shape Retrieval", "Cross Modal Guidance Network", "Three Dimensional Displays", "Shape", "Two Dimensional Displays", "Training", "Benchmark Testing", "Task Analysis", "Feature Extraction", "Sketch", "3 D Shape Retrieval", "Cross Modal Differences", "Guidance Network", "Feature Alignment" ], "authors": [ { "affiliation": "Tongji University,School of Software Engineering,China", "fullName": "Weidong Dai", "givenName": "Weidong", "surname": "Dai", "__typename": "ArticleAuthorType" }, { "affiliation": "Tongji University,School of Software Engineering,China", "fullName": "Shuang Liang", "givenName": "Shuang", "surname": "Liang", "__typename": "ArticleAuthorType" } ], "idPrefix": "icme", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-07-01T00:00:00", "pubType": "proceedings", "pages": "1-6", "year": "2020", "issn": null, "isbn": "978-1-7281-1331-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09102908", "articleId": "1kwqSoV3HP2", "__typename": "AdjacentArticleType" }, "next": { "fno": "09102867", "articleId": "1kwrafKJPYk", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2017/0457/0/0457d615", "title": "Learning Barycentric Representations of 3D Shapes for Sketch-Based 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457d615/12OmNB0FxiX", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2017/6067/0/08019464", "title": "Multi-view pairwise relationship learning for sketch based 3D shape retrieval", "doi": null, "abstractUrl": "/proceedings-article/icme/2017/08019464/12OmNy6Zs2q", "parentPublication": { "id": "proceedings/icme/2017/6067/0", "title": "2017 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000a089", "title": "Cross-Modal Deep Variational Hand Pose Estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000a089/17D45Xh13pi", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/08/09007505", "title": "Sketch Augmentation-Driven Shape Retrieval Learning Framework Based on Convolutional Neural Networks", "doi": null, "abstractUrl": "/journal/tg/2021/08/09007505/1hJKlMJzueI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2020/6553/0/09093316", "title": "3D Hand Pose Estimation with Disentangled Cross-Modal Latent Space", "doi": null, "abstractUrl": "/proceedings-article/wacv/2020/09093316/1jPbFBfZZAI", "parentPublication": { "id": "proceedings/wacv/2020/6553/0", "title": "2020 IEEE Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800m2602", "title": "xMUDA: Cross-Modal Unsupervised Domain Adaptation for 3D Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800m2602/1m3oihsayQg", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a081", "title": "Towards 3D VR-Sketch to 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a081/1qyxlDtR0Ji", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2020/9234/0/923400a184", "title": "Deep 3D Shape Reconstruction from Single-View Sketch Image", "doi": null, "abstractUrl": "/proceedings-article/icdh/2020/923400a184/1uGY2GTiIda", "parentPublication": { "id": "proceedings/icdh/2020/9234/0", "title": "2020 8th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900d141", "title": "Cross-Modal Center Loss for 3D Cross-Modal Retrieval", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900d141/1yeJANS7HfG", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900l1789", "title": "PointAugmenting: Cross-Modal Augmentation for 3D Object Detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900l1789/1yeJNb0uIg0", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1nHRQncZfOM", "title": "2019 International Conference on Virtual Reality and Visualization (ICVRV)", "acronym": "icvrv", "groupId": "1800579", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1nHRUrDMgE0", "doi": "10.1109/ICVRV47840.2019.00026", "title": "Automatic Hair Modeling from One Image", "normalizedTitle": "Automatic Hair Modeling from One Image", "abstract": "Hair is one of the most critical characteristics of a person in the process of digitizing characters, but at the same time, hair modeling is still a very challenging task due to the diversity of hairstyles and the overlap among hair. We introduce a method to automatically generate 3D hair geometry from a front hair image, which can recover the outline and details of the hair geometry. We designed a encoder-decoder convolutional neural network which takes the 2D orientation field from a hair image as input, and output the characteristic hair geometry. Then we use the characteristic strands to search for eligible hairstyle from Hairstyle-database, and fuse retrieved hair model to get the nal hairstyle. This pipeline can automatically recover hair geometry from a front hair image without any supplementary information. Experimental results show that our approach achieves realistic reconstruction effect from real Internet pictures and self-portraits.", "abstracts": [ { "abstractType": "Regular", "content": "Hair is one of the most critical characteristics of a person in the process of digitizing characters, but at the same time, hair modeling is still a very challenging task due to the diversity of hairstyles and the overlap among hair. We introduce a method to automatically generate 3D hair geometry from a front hair image, which can recover the outline and details of the hair geometry. We designed a encoder-decoder convolutional neural network which takes the 2D orientation field from a hair image as input, and output the characteristic hair geometry. Then we use the characteristic strands to search for eligible hairstyle from Hairstyle-database, and fuse retrieved hair model to get the nal hairstyle. This pipeline can automatically recover hair geometry from a front hair image without any supplementary information. Experimental results show that our approach achieves realistic reconstruction effect from real Internet pictures and self-portraits.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Hair is one of the most critical characteristics of a person in the process of digitizing characters, but at the same time, hair modeling is still a very challenging task due to the diversity of hairstyles and the overlap among hair. We introduce a method to automatically generate 3D hair geometry from a front hair image, which can recover the outline and details of the hair geometry. We designed a encoder-decoder convolutional neural network which takes the 2D orientation field from a hair image as input, and output the characteristic hair geometry. Then we use the characteristic strands to search for eligible hairstyle from Hairstyle-database, and fuse retrieved hair model to get the nal hairstyle. This pipeline can automatically recover hair geometry from a front hair image without any supplementary information. Experimental results show that our approach achieves realistic reconstruction effect from real Internet pictures and self-portraits.", "fno": "09212824", "keywords": [ "Computational Geometry", "Convolutional Neural Nets", "Feature Extraction", "Image Fusion", "Image Reconstruction", "Realistic Images", "Solid Modelling", "Automatic Hair Modeling", "Hair Image", "3 D Hair Geometry", "Encoder Decoder Convolutional Neural Network", "2 D Orientation Field", "Hairstyle Database", "Hair Model Fusion", "Realistic Reconstruction", "Hair", "Two Dimensional Displays", "Three Dimensional Displays", "Solid Modeling", "Databases", "Image Reconstruction", "Training", "Single View Hair Modeling", "Deep Learning", "Hairstyles Database", "Orientation Field" ], "authors": [ { "affiliation": "Beihang University , Qingdao , China", "fullName": "Ligang Cheng", "givenName": "Ligang", "surname": "Cheng", "__typename": "ArticleAuthorType" }, { "affiliation": "Shandong University of Science and Technology, Qingdao , China", "fullName": "Yongtang Bao", "givenName": "Yongtang", "surname": "Bao", "__typename": "ArticleAuthorType" }, { "affiliation": "Beihang University, Beijing, China", "fullName": "Yue Qi", "givenName": "Yue", "surname": "Qi", "__typename": "ArticleAuthorType" } ], "idPrefix": "icvrv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-11-01T00:00:00", "pubType": "proceedings", "pages": "108-112", "year": "2019", "issn": "2375-141X", "isbn": "978-1-7281-4752-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09212946", "articleId": "1nHRSx20hQA", "__typename": "AdjacentArticleType" }, "next": { "fno": "09212976", "articleId": "1nHRSEYLVEA", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cad-cg/2005/2473/0/24730489", "title": "Scattering-Based Interactive Hair Rendering", "doi": null, "abstractUrl": "/proceedings-article/cad-cg/2005/24730489/12OmNvjQ8Ct", "parentPublication": { "id": "proceedings/cad-cg/2005/2473/0", "title": "Ninth International Conference on Computer Aided Design and Computer Graphics (CAD-CG'05)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2012/1226/0/188P2A38", "title": "Multi-view hair capture using orientation fields", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2012/188P2A38/12OmNxV4iu4", "parentPublication": { "id": "proceedings/cvpr/2012/1226/0", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2017/2089/0/2089a214", "title": "NPR Hair Modeling with Parametric Clumps", "doi": null, "abstractUrl": "/proceedings-article/cw/2017/2089a214/12OmNxuo0jO", "parentPublication": { "id": "proceedings/cw/2017/2089/0", "title": "2017 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2015/7568/0/7568a411", "title": "Image-Based Hair Pre-processing for Art Creation: A Case Study of Bas-Relief Modelling", "doi": null, "abstractUrl": "/proceedings-article/iv/2015/7568a411/12OmNzBwGJv", "parentPublication": { "id": "proceedings/iv/2015/7568/0", "title": "2015 19th International Conference on Information Visualisation (iV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600b516", "title": "NeuralHDHair: Automatic High-fidelity Hair Modeling from a Single Image Using Implicit Neural Representations", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600b516/1H1lkq5sTPq", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2018/8497/0/849700a008", "title": "Data-Driven Hair Modeling from a Single Image", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2018/849700a008/1a3x7jLsFPi", "parentPublication": { "id": "proceedings/icvrv/2018/8497/0", "title": "2018 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08964443", "title": "DeepSketchHair: Deep Sketch-Based 3D Hair Modeling", "doi": null, "abstractUrl": "/journal/tg/2021/07/08964443/1gLZSnCp3Ko", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800h707", "title": "Deep 3D Portrait From a Single Image", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800h707/1m3nijIcYta", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800h444", "title": "Intuitive, Interactive Beard and Hair Synthesis With Generative Models", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800h444/1m3ol6XnJ3q", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900m2798", "title": "i3DMM: Deep Implicit 3D Morphable Model of Human Heads", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900m2798/1yeLR7aJqiA", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1qyxi3OgORy", "title": "2020 International Conference on 3D Vision (3DV)", "acronym": "3dv", "groupId": "1800494", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1qyxlDtR0Ji", "doi": "10.1109/3DV50981.2020.00018", "title": "Towards 3D VR-Sketch to 3D Shape Retrieval", "normalizedTitle": "Towards 3D VR-Sketch to 3D Shape Retrieval", "abstract": "Growing free online 3D shapes collections dictated research on 3D retrieval. Active debate has however been had on (i) what the best input modality is to trigger retrieval, and (ii) the ultimate usage scenario for such retrieval. In this paper, we offer a different perspective towards answering these questions - we study the use of 3D sketches as an input modality and advocate a VR-scenario where retrieval is conducted. Thus, the ultimate vision is that users can freely retrieve a 3D model by air-doodling in a VR environment. As a first stab at this new 3D VR-sketch to 3D shape retrieval problem, we make four contributions. First, we code a VR utility to collect 3D VR-sketches and conduct retrieval. Second, we collect the first set of 167 3D VRsketches on two shape categories from ModelNet. Third, we propose a novel approach to generate a synthetic dataset of human-like 3D sketches of different abstract levels to train deep networks. At last, we compare the common multi-view and volumetric approaches: We show that, in contrast to 3D shape to 3D shape retrieval, volumetric point-based approaches exhibit superior performance on 3D sketch to 3D shape retrieval due to the sparse and abstract nature of 3D VR-sketches. We believe these contributions will collectively serve as enablers for future attempts at this problem. The VR interface, code and datasets are available at https://tinyurl.com/3DSketch3DV.", "abstracts": [ { "abstractType": "Regular", "content": "Growing free online 3D shapes collections dictated research on 3D retrieval. Active debate has however been had on (i) what the best input modality is to trigger retrieval, and (ii) the ultimate usage scenario for such retrieval. In this paper, we offer a different perspective towards answering these questions - we study the use of 3D sketches as an input modality and advocate a VR-scenario where retrieval is conducted. Thus, the ultimate vision is that users can freely retrieve a 3D model by air-doodling in a VR environment. As a first stab at this new 3D VR-sketch to 3D shape retrieval problem, we make four contributions. First, we code a VR utility to collect 3D VR-sketches and conduct retrieval. Second, we collect the first set of 167 3D VRsketches on two shape categories from ModelNet. Third, we propose a novel approach to generate a synthetic dataset of human-like 3D sketches of different abstract levels to train deep networks. At last, we compare the common multi-view and volumetric approaches: We show that, in contrast to 3D shape to 3D shape retrieval, volumetric point-based approaches exhibit superior performance on 3D sketch to 3D shape retrieval due to the sparse and abstract nature of 3D VR-sketches. We believe these contributions will collectively serve as enablers for future attempts at this problem. The VR interface, code and datasets are available at https://tinyurl.com/3DSketch3DV.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Growing free online 3D shapes collections dictated research on 3D retrieval. Active debate has however been had on (i) what the best input modality is to trigger retrieval, and (ii) the ultimate usage scenario for such retrieval. In this paper, we offer a different perspective towards answering these questions - we study the use of 3D sketches as an input modality and advocate a VR-scenario where retrieval is conducted. Thus, the ultimate vision is that users can freely retrieve a 3D model by air-doodling in a VR environment. As a first stab at this new 3D VR-sketch to 3D shape retrieval problem, we make four contributions. First, we code a VR utility to collect 3D VR-sketches and conduct retrieval. Second, we collect the first set of 167 3D VRsketches on two shape categories from ModelNet. Third, we propose a novel approach to generate a synthetic dataset of human-like 3D sketches of different abstract levels to train deep networks. At last, we compare the common multi-view and volumetric approaches: We show that, in contrast to 3D shape to 3D shape retrieval, volumetric point-based approaches exhibit superior performance on 3D sketch to 3D shape retrieval due to the sparse and abstract nature of 3D VR-sketches. We believe these contributions will collectively serve as enablers for future attempts at this problem. The VR interface, code and datasets are available at https://tinyurl.com/3DSketch3DV.", "fno": "812800a081", "keywords": [ "Image Retrieval", "Solid Modelling", "Virtual Reality", "VR Environment", "3 D VR Sketches", "Conduct Retrieval", "3 D Shape Retrieval", "Model Net", "Three Dimensional Displays", "Shape", "Solid Modeling", "Two Dimensional Displays", "Task Analysis", "Training Data", "Load Modeling", "VR 3 D Sketch", "3 D Sketch Based Retrieval", "Triplet Loss", "Triplet Center Loss", "3 D VR Sketch Dataset", "3 D Sketch NPR", "Abstract 3 D Sketch" ], "authors": [ { "affiliation": "SketchX, CVSSP, University of Surrey", "fullName": "Ling Luo", "givenName": "Ling", "surname": "Luo", "__typename": "ArticleAuthorType" }, { "affiliation": "SketchX, CVSSP, University of Surrey", "fullName": "Yulia Gryaditskaya", "givenName": "Yulia", "surname": "Gryaditskaya", "__typename": "ArticleAuthorType" }, { "affiliation": "SketchX, CVSSP, University of Surrey", "fullName": "Yongxin Yang", "givenName": "Yongxin", "surname": "Yang", "__typename": "ArticleAuthorType" }, { "affiliation": "SketchX, CVSSP, University of Surrey", "fullName": "Tao Xiang", "givenName": "Tao", "surname": "Xiang", "__typename": "ArticleAuthorType" }, { "affiliation": "SketchX, CVSSP, University of Surrey", "fullName": "Yi-Zhe Song", "givenName": "Yi-Zhe", "surname": "Song", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-11-01T00:00:00", "pubType": "proceedings", "pages": "81-90", "year": "2020", "issn": null, "isbn": "978-1-7281-8128-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "812800a071", "articleId": "1qyxmgpJORW", "__typename": "AdjacentArticleType" }, "next": { "fno": "812800a091", "articleId": "1qyxpav5shG", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2017/0457/0/0457d615", "title": "Learning Barycentric Representations of 3D Shapes for Sketch-Based 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457d615/12OmNB0FxiX", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2017/6067/0/08019464", "title": "Multi-view pairwise relationship learning for sketch based 3D shape retrieval", "doi": null, "abstractUrl": "/proceedings-article/icme/2017/08019464/12OmNy6Zs2q", "parentPublication": { "id": "proceedings/icme/2017/6067/0", "title": "2017 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2017/06/mcg2017060088", "title": "Sketch-Based Articulated 3D Shape Retrieval", "doi": null, "abstractUrl": "/magazine/cg/2017/06/mcg2017060088/13rRUwfqpG7", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2018/9497/0/949700a311", "title": "Sketch-Based Shape Retrieval via Multi-view Attention and Generalized Similarity", "doi": null, "abstractUrl": "/proceedings-article/icdh/2018/949700a311/17D45VObpQZ", "parentPublication": { "id": "proceedings/icdh/2018/9497/0", "title": "2018 7th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2022/5670/0/567000a383", "title": "Structure-Aware 3D VR Sketch to 3D Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/3dv/2022/567000a383/1KYsqgmUniE", "parentPublication": { "id": "proceedings/3dv/2022/5670/0", "title": "2022 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2016/4847/0/07900083", "title": "3D sketch-based 3D model retrieval with convolutional neural network", "doi": null, "abstractUrl": "/proceedings-article/icpr/2016/07900083/1gysq8EnfHi", "parentPublication": { "id": "proceedings/icpr/2016/4847/0", "title": "2016 23rd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/08/09007505", "title": "Sketch Augmentation-Driven Shape Retrieval Learning Framework Based on Convolutional Neural Networks", "doi": null, "abstractUrl": "/journal/tg/2021/08/09007505/1hJKlMJzueI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2020/1331/0/09102925", "title": "Cross-Modal Guidance Network For Sketch-Based 3d Shape Retrieval", "doi": null, "abstractUrl": "/proceedings-article/icme/2020/09102925/1kwqTrDSXF6", "parentPublication": { "id": "proceedings/icme/2020/1331/0", "title": "2020 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2020/0497/0/049700a005", "title": "Sketch-based 3D shape retrieval via attention", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2020/049700a005/1vg7Y9U3P5S", "parentPublication": { "id": "proceedings/icvrv/2020/0497/0", "title": "2020 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icicta/2020/8666/0/866600a223", "title": "Sketch-based 3D Shape Retrieval with Multi-Silhouette View Based on Convolutional Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icicta/2020/866600a223/1wRIvGNgH9m", "parentPublication": { "id": "proceedings/icicta/2020/8666/0", "title": "2020 13th International Conference on Intelligent Computation Technology and Automation (ICICTA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNCmpcNk", "title": "Visualization Conference, IEEE", "acronym": "ieee-vis", "groupId": "1000796", "volume": "0", "displayVolume": "0", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "12OmNvpewaO", "doi": "10.1109/VIS.2005.128", "title": "Visualizing Tensor Fields in Geomechanics", "normalizedTitle": "Visualizing Tensor Fields in Geomechanics", "abstract": "The study of stress and strains in soils and structures (solids) help us gain a better understanding of events such as failure of bridges, dams and buildings, or accumulated stresses and strains in geological subduction zones that could trigger earthquakes and subsequently tsunamis. In such domains, the key feature of interest is the location and orientation of maximal shearing planes. This paper describes a method that highlights this feature in stress tensor fields. It uses a plane-in-a-box glyph which provides a global perspective of shearing planes based on local analysis of tensors. The analysis can be performed over the entire domain, or the user can interactively specify where to introduce these glyphs. Alternatively, they can also be placed depending on the threshold level of several physical relevant parameters such as double couple and compensated linear vector dipole. Both methods are tested on stress tensor fields from geomechanics.", "abstracts": [ { "abstractType": "Regular", "content": "The study of stress and strains in soils and structures (solids) help us gain a better understanding of events such as failure of bridges, dams and buildings, or accumulated stresses and strains in geological subduction zones that could trigger earthquakes and subsequently tsunamis. In such domains, the key feature of interest is the location and orientation of maximal shearing planes. This paper describes a method that highlights this feature in stress tensor fields. It uses a plane-in-a-box glyph which provides a global perspective of shearing planes based on local analysis of tensors. The analysis can be performed over the entire domain, or the user can interactively specify where to introduce these glyphs. Alternatively, they can also be placed depending on the threshold level of several physical relevant parameters such as double couple and compensated linear vector dipole. Both methods are tested on stress tensor fields from geomechanics.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The study of stress and strains in soils and structures (solids) help us gain a better understanding of events such as failure of bridges, dams and buildings, or accumulated stresses and strains in geological subduction zones that could trigger earthquakes and subsequently tsunamis. In such domains, the key feature of interest is the location and orientation of maximal shearing planes. This paper describes a method that highlights this feature in stress tensor fields. It uses a plane-in-a-box glyph which provides a global perspective of shearing planes based on local analysis of tensors. The analysis can be performed over the entire domain, or the user can interactively specify where to introduce these glyphs. Alternatively, they can also be placed depending on the threshold level of several physical relevant parameters such as double couple and compensated linear vector dipole. Both methods are tested on stress tensor fields from geomechanics.", "fno": "27660005", "keywords": [ "Symmetric Tensors", "Stress Tensor", "Seismic Moment Tensor", "Anisotropic", "Deviatoric", "Double Couple", "Compensated Linear Vector Dipole" ], "authors": [ { "affiliation": "UCSC", "fullName": "Alisa Neeman", "givenName": "Alisa", "surname": "Neeman", "__typename": "ArticleAuthorType" }, { "affiliation": "UC Davis", "fullName": "Boris Jeremic", "givenName": "Boris", "surname": "Jeremic", "__typename": "ArticleAuthorType" }, { "affiliation": "UCSC", "fullName": "Alex Pang", "givenName": "Alex", "surname": "Pang", "__typename": "ArticleAuthorType" } ], "idPrefix": "ieee-vis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2005-10-01T00:00:00", "pubType": "proceedings", "pages": "5", "year": "2005", "issn": null, "isbn": "0-7803-9462-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "01532790", "articleId": "12OmNxxdZyx", "__typename": "AdjacentArticleType" }, "next": { "fno": "01532791", "articleId": "12OmNyQpgSW", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ieee-vis/2004/8788/0/87880313", "title": "Topological Lines in 3D Tensor Fields", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2004/87880313/12OmNApLGKA", "parentPublication": { "id": "proceedings/ieee-vis/2004/8788/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/27660002", "title": "Exploring 2D Tensor Fields Using Stress Nets", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660002/12OmNBCZnRL", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/27660004", "title": "HOT- Lines: Tracking Lines in Higher Order Tensor Fields", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660004/12OmNwMXnqd", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/1998/9176/0/91760297", "title": "Interactive Deformations from Tensor Fields", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/1998/91760297/12OmNyQGSpp", "parentPublication": { "id": "proceedings/ieee-vis/1998/9176/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/01532774", "title": "Visualizing tensor fields in geomechanics", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/01532774/12OmNyk300u", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/1997/8262/0/82620059", "title": "Singularities in nonuniform tensor fields", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/1997/82620059/12OmNzFMFiX", "parentPublication": { "id": "proceedings/ieee-vis/1997/8262/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2004/8788/0/87880123", "title": "Physically Based Methods for Tensor Field Visualization", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2004/87880123/12OmNzTppFk", "parentPublication": { "id": "proceedings/ieee-vis/2004/8788/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2008/06/ttg2008061627", "title": "Invariant Crease Lines for Topological and Structural Analysis of Tensor Fields", "doi": null, "abstractUrl": "/journal/tg/2008/06/ttg2008061627/13rRUxASu0F", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2011/06/ttg2011060781", "title": "Coherent Structures of Characteristic Curves in Symmetric Second Order Tensor Fields", "doi": null, "abstractUrl": "/journal/tg/2011/06/ttg2011060781/13rRUyv53Fl", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/1993/04/mcg1993040025", "title": "Visualizing Second-Order Tensor Fields with Hyperstreamlines", "doi": null, "abstractUrl": "/magazine/cg/1993/04/mcg1993040025/13rRUyv53HF", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxWuirg", "title": "Visualization Conference, IEEE", "acronym": "ieee-vis", "groupId": "1000796", "volume": "0", "displayVolume": "0", "year": "2004", "__typename": "ProceedingType" }, "article": { "id": "12OmNvqEvJq", "doi": "10.1109/VISUAL.2004.115", "title": "Visualization of Salt-Induced Stress Perturbations", "normalizedTitle": "Visualization of Salt-Induced Stress Perturbations", "abstract": "An important challenge encountered during post-processing of finite element analyses is the visualizing of three-dimensional fields of real-valued second-order tensors. Namely, as finite element meshes become more complex and detailed, evaluation and presentation of the principal stresses becomes correspondingly problematic. In this paper, we describe techniques used to visualize simulations of perturbed in-situ stress fields associated with hypothetical salt bodies in the Gulf of Mexico. We present an adaptation of the Mohr diagram, a graphical paper and pencil method used by the material mechanics community for estimating coordinate transformations for stress tensors, as a new tensor glyph for dynamically exploring tensor variables within three-dimensional finite element models. This interactive glyph can be used as either a probe or a filter through brushing and linking.", "abstracts": [ { "abstractType": "Regular", "content": "An important challenge encountered during post-processing of finite element analyses is the visualizing of three-dimensional fields of real-valued second-order tensors. Namely, as finite element meshes become more complex and detailed, evaluation and presentation of the principal stresses becomes correspondingly problematic. In this paper, we describe techniques used to visualize simulations of perturbed in-situ stress fields associated with hypothetical salt bodies in the Gulf of Mexico. We present an adaptation of the Mohr diagram, a graphical paper and pencil method used by the material mechanics community for estimating coordinate transformations for stress tensors, as a new tensor glyph for dynamically exploring tensor variables within three-dimensional finite element models. This interactive glyph can be used as either a probe or a filter through brushing and linking.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "An important challenge encountered during post-processing of finite element analyses is the visualizing of three-dimensional fields of real-valued second-order tensors. Namely, as finite element meshes become more complex and detailed, evaluation and presentation of the principal stresses becomes correspondingly problematic. In this paper, we describe techniques used to visualize simulations of perturbed in-situ stress fields associated with hypothetical salt bodies in the Gulf of Mexico. We present an adaptation of the Mohr diagram, a graphical paper and pencil method used by the material mechanics community for estimating coordinate transformations for stress tensors, as a new tensor glyph for dynamically exploring tensor variables within three-dimensional finite element models. This interactive glyph can be used as either a probe or a filter through brushing and linking.", "fno": "87880369", "keywords": [ "Tensor Field Visualization", "Mohrs Circles", "Visual Debugging", "Finite Element Codes And Simulations" ], "authors": [ { "affiliation": "Sandia National Laboratories", "fullName": "Patricia Crossno", "givenName": "Patricia", "surname": "Crossno", "__typename": "ArticleAuthorType" }, { "affiliation": "Sandia National Laboratories", "fullName": "David H. Rogers", "givenName": "David H.", "surname": "Rogers", "__typename": "ArticleAuthorType" }, { "affiliation": "Sandia National Laboratories", "fullName": "Rebecca M. Brannon", "givenName": "Rebecca M.", "surname": "Brannon", "__typename": "ArticleAuthorType" }, { "affiliation": "Los Alamos National Laboratories", "fullName": "David Coblentz", "givenName": "David", "surname": "Coblentz", "__typename": "ArticleAuthorType" } ], "idPrefix": "ieee-vis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2004-10-01T00:00:00", "pubType": "proceedings", "pages": "369-376", "year": "2004", "issn": null, "isbn": "0-7803-8788-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "87880361", "articleId": "12OmNy1SFMH", "__typename": "AdjacentArticleType" }, "next": { "fno": "87880377", "articleId": "12OmNvC0sXh", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ieee-vis/2005/2766/0/27660002", "title": "Exploring 2D Tensor Fields Using Stress Nets", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660002/12OmNBCZnRL", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/visual/1991/2245/0/00175823", "title": "Applying 3D visualization techniques to finite element analysis", "doi": null, "abstractUrl": "/proceedings-article/visual/1991/00175823/12OmNqBKTSD", "parentPublication": { "id": "proceedings/visual/1991/2245/0", "title": "1991 Proceeding Visualization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/27660005", "title": "Visualizing Tensor Fields in Geomechanics", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660005/12OmNvpewaO", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2002/7498/0/7498sigfrids", "title": "Tensor Field Visualisation using Adaptive Filtering of Noise Fields combined with Glyph Rendering", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2002/7498sigfrids/12OmNyfdOXA", "parentPublication": { "id": "proceedings/ieee-vis/2002/7498/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2004/8788/0/87880123", "title": "Physically Based Methods for Tensor Field Visualization", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2004/87880123/12OmNzTppFk", "parentPublication": { "id": "proceedings/ieee-vis/2004/8788/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/01/07192722", "title": "Glyph-Based Comparative Visualization for Diffusion Tensor Fields", "doi": null, "abstractUrl": "/journal/tg/2016/01/07192722/13rRUx0gefn", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2006/05/v1329", "title": "Diffusion Tensor Visualization with Glyph Packing", "doi": null, "abstractUrl": "/journal/tg/2006/05/v1329/13rRUxYrbUu", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2010/06/ttg2010061595", "title": "Superquadric Glyphs for Symmetric Second-Order Tensors", "doi": null, "abstractUrl": "/journal/tg/2010/06/ttg2010061595/13rRUxZzAhA", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2005/05/v0508", "title": "Visualization of Geologic Stress Perturbations Using Mohr Diagrams", "doi": null, "abstractUrl": "/journal/tg/2005/05/v0508/13rRUyeTVhT", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08967163", "title": "Visualization of 3D Stress Tensor Fields Using Superquadric Glyphs on Displacement Streamlines", "doi": null, "abstractUrl": "/journal/tg/2021/07/08967163/1gPjyn904OA", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNCmpcNk", "title": "Visualization Conference, IEEE", "acronym": "ieee-vis", "groupId": "1000796", "volume": "0", "displayVolume": "0", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "12OmNzmtWye", "doi": "10.1109/VISUAL.2005.1532771", "title": "Exploring 2D tensor fields using stress nets", "normalizedTitle": "Exploring 2D tensor fields using stress nets", "abstract": "In this article we describe stress nets, a technique for exploring 2D tensor fields. Our method allows a user to examine simultaneously the tensors' eigenvectors (both major and minor) as well as scalar-valued tensor invariants. By avoiding noise-advection techniques, we are able to display both principal directions of the tensor field as well as the derived scalars without cluttering the display. We present a CPU-only implementation of stress nets as well as a hybrid CPU/GPU approach and discuss the relative strengths and weaknesses of each. Stress nets have been used as part of an investigation into crack propagation. They were used to display the directions of maximum shear in a slab of material under tension as well as the magnitude of the shear forces acting on each point. Our methods allowed users to find new features in the data that were not visible on standard plots of tensor invariants. These features disagree with commonly accepted analytical crack propagation solutions and have sparked renewed investigation. Though developed for a materials mechanics problem, our method applies equally well to any 2D tensor field having unique characteristic directions.", "abstracts": [ { "abstractType": "Regular", "content": "In this article we describe stress nets, a technique for exploring 2D tensor fields. Our method allows a user to examine simultaneously the tensors' eigenvectors (both major and minor) as well as scalar-valued tensor invariants. By avoiding noise-advection techniques, we are able to display both principal directions of the tensor field as well as the derived scalars without cluttering the display. We present a CPU-only implementation of stress nets as well as a hybrid CPU/GPU approach and discuss the relative strengths and weaknesses of each. Stress nets have been used as part of an investigation into crack propagation. They were used to display the directions of maximum shear in a slab of material under tension as well as the magnitude of the shear forces acting on each point. Our methods allowed users to find new features in the data that were not visible on standard plots of tensor invariants. These features disagree with commonly accepted analytical crack propagation solutions and have sparked renewed investigation. Though developed for a materials mechanics problem, our method applies equally well to any 2D tensor field having unique characteristic directions.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this article we describe stress nets, a technique for exploring 2D tensor fields. Our method allows a user to examine simultaneously the tensors' eigenvectors (both major and minor) as well as scalar-valued tensor invariants. By avoiding noise-advection techniques, we are able to display both principal directions of the tensor field as well as the derived scalars without cluttering the display. We present a CPU-only implementation of stress nets as well as a hybrid CPU/GPU approach and discuss the relative strengths and weaknesses of each. Stress nets have been used as part of an investigation into crack propagation. They were used to display the directions of maximum shear in a slab of material under tension as well as the magnitude of the shear forces acting on each point. Our methods allowed users to find new features in the data that were not visible on standard plots of tensor invariants. These features disagree with commonly accepted analytical crack propagation solutions and have sparked renewed investigation. Though developed for a materials mechanics problem, our method applies equally well to any 2D tensor field having unique characteristic directions.", "fno": "01532771", "keywords": [ "Data Visualisation", "Physics Computing", "Stress Analysis", "Cracks", "Tensors", "2 D Tensor Field", "Stress Nets", "Noise Advection Techniques", "Hybrid CPU Approach", "Hybrid GPU Approach", "Crack Propagation", "Shear Forces", "Tensile Stress", "Displays", "Data Visualization", "Laboratories", "Streaming Media", "Capacitive Sensors", "Data Mining", "Occupational Stress", "Slabs", "Chromium" ], "authors": [ { "affiliation": "Sandia Nat. Labs., Albuquerque, NM, USA", "fullName": "A. Wilson", "givenName": "A.", "surname": "Wilson", "__typename": "ArticleAuthorType" }, { "affiliation": "Sandia Nat. Labs., Albuquerque, NM, USA", "fullName": "R. Brannon", "givenName": "R.", "surname": "Brannon", "__typename": "ArticleAuthorType" } ], "idPrefix": "ieee-vis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2005-01-01T00:00:00", "pubType": "proceedings", "pages": "11,12,13,14,15,16,17,18", "year": "2005", "issn": null, "isbn": null, "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "01532770", "articleId": "12OmNCw3z9K", "__typename": "AdjacentArticleType" }, "next": { "fno": "01532772", "articleId": "12OmNyo1o3L", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ieee-vis/2005/2766/0/27660002", "title": "Exploring 2D Tensor Fields Using Stress Nets", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660002/12OmNBCZnRL", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/01532770", "title": "2D asymmetric tensor analysis", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/01532770/12OmNCw3z9K", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/27660005", "title": "Visualizing Tensor Fields in Geomechanics", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660005/12OmNvpewaO", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/01532841", "title": "Topological structures of 3D tensor fields", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/01532841/12OmNx5GTXp", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/1998/9176/0/91760297", "title": "Interactive Deformations from Tensor Fields", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/1998/91760297/12OmNyQGSpp", "parentPublication": { "id": "proceedings/ieee-vis/1998/9176/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/01532774", "title": "Visualizing tensor fields in geomechanics", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/01532774/12OmNyk300u", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/1997/8262/0/82620059", "title": "Singularities in nonuniform tensor fields", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/1997/82620059/12OmNzFMFiX", "parentPublication": { "id": "proceedings/ieee-vis/1997/8262/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2009/06/ttg2009061399", "title": "Stress Tensor Field Visualization for Implant Planning in Orthopedics", "doi": null, "abstractUrl": "/journal/tg/2009/06/ttg2009061399/13rRUwI5U2B", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/1993/04/mcg1993040025", "title": "Visualizing Second-Order Tensor Fields with Hyperstreamlines", "doi": null, "abstractUrl": "/magazine/cg/1993/04/mcg1993040025/13rRUyv53HF", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08967163", "title": "Visualization of 3D Stress Tensor Fields Using Superquadric Glyphs on Displacement Streamlines", "doi": null, "abstractUrl": "/journal/tg/2021/07/08967163/1gPjyn904OA", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvAiSpZ", "title": "2015 IEEE Virtual Reality (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNB1NVNQ", "doi": "10.1109/VR.2015.7223431", "title": "Walking recording and experience system by Visual Psychophysics Lab", "normalizedTitle": "Walking recording and experience system by Visual Psychophysics Lab", "abstract": "We aim to develop a virtual-reality system that records a person's walking experience and gives other users the experience of his/her walking. We recorded stereo motion images of video cameras on a person's forehead with synchronous acceleration data of ankles. Then, we presented stereo motion images on a HMD with synchronous vibrations on soles of observer's feet. Observers reported better experience of vection, walking, and tele-existence from stereo images with vibrations than without vibrations. We recorded walking experiences of different body sizes including a child (130 cm tall) and those of a dog. Observers can partly experience child's walking and even dog's running.", "abstracts": [ { "abstractType": "Regular", "content": "We aim to develop a virtual-reality system that records a person's walking experience and gives other users the experience of his/her walking. We recorded stereo motion images of video cameras on a person's forehead with synchronous acceleration data of ankles. Then, we presented stereo motion images on a HMD with synchronous vibrations on soles of observer's feet. Observers reported better experience of vection, walking, and tele-existence from stereo images with vibrations than without vibrations. We recorded walking experiences of different body sizes including a child (130 cm tall) and those of a dog. Observers can partly experience child's walking and even dog's running.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We aim to develop a virtual-reality system that records a person's walking experience and gives other users the experience of his/her walking. We recorded stereo motion images of video cameras on a person's forehead with synchronous acceleration data of ankles. Then, we presented stereo motion images on a HMD with synchronous vibrations on soles of observer's feet. Observers reported better experience of vection, walking, and tele-existence from stereo images with vibrations than without vibrations. We recorded walking experiences of different body sizes including a child (130 cm tall) and those of a dog. Observers can partly experience child's walking and even dog's running.", "fno": "07223431", "keywords": [ "Legged Locomotion", "Visualization", "Vibrations", "Observers", "Virtual Reality", "Foot", "Acceleration", "Stereo Vision", "Vection", "Vibration", "Walking" ], "authors": [ { "affiliation": "Graduate School of Engineering, Toyohashi University of Technology", "fullName": "Atsuhiro Fujita", "givenName": "Atsuhiro", "surname": "Fujita", "__typename": "ArticleAuthorType" }, { "affiliation": "Graduate School of Engineering, Toyohashi University of Technology", "fullName": "Shohei Ueda", "givenName": "Shohei", "surname": "Ueda", "__typename": "ArticleAuthorType" }, { "affiliation": "Graduate School of Engineering, Toyohashi University of Technology", "fullName": "Junki Nozawa", "givenName": "Junki", "surname": "Nozawa", "__typename": "ArticleAuthorType" }, { "affiliation": "Interfaculty Initiative in Information Studies, The University of Tokyo", "fullName": "Koichi Hirota", "givenName": "Koichi", "surname": "Hirota", "__typename": "ArticleAuthorType" }, { "affiliation": "Faculty of System Design, Tokyo Metropolitan University", "fullName": "Yasushi Ikei", "givenName": "Yasushi", "surname": "Ikei", "__typename": "ArticleAuthorType" }, { "affiliation": "Department of Computer Science and Engineering, Toyohashi University of Technology", "fullName": "Michiteru Kitazaki", "givenName": "Michiteru", "surname": "Kitazaki", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-03-01T00:00:00", "pubType": "proceedings", "pages": "333-334", "year": "2015", "issn": null, "isbn": "978-1-4799-1727-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07223430", "articleId": "12OmNviHKdt", "__typename": "AdjacentArticleType" }, "next": { "fno": "07223432", "articleId": "12OmNB9bvqr", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dui/2016/0842/0/07460066", "title": "Rhythmic vibrations to heels and forefeet to produce virtual walking", "doi": null, "abstractUrl": "/proceedings-article/3dui/2016/07460066/12OmNBQkwZJ", "parentPublication": { "id": "proceedings/3dui/2016/0842/0", "title": "2016 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2016/0836/0/07504709", "title": "The effect of multi-sensory cues on performance and experience during walking in immersive virtual environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504709/12OmNyrqzC0", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08448288", "title": "Experiencing an Invisible World War I Battlefield Through Narrative-Driven Redirected Walking in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08448288/13bd1fZBGdu", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ta/2017/03/07450621", "title": "Emotion Rendering in Plantar Vibro-Tactile Simulations of Imagined Walking Styles", "doi": null, "abstractUrl": "/journal/ta/2017/03/07450621/13rRUwIF6cq", "parentPublication": { "id": "trans/ta", "title": "IEEE Transactions on Affective Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09911682", "title": "Effect of Vibrations on Impression of Walking and Embodiment With First- and Third-Person Avatar", "doi": null, "abstractUrl": "/journal/tg/5555/01/09911682/1HeiWQWKlTG", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090634", "title": "Rhythmic proprioceptive stimulation improves embodiment in a walking avatar when added to visual stimulation", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090634/1jIxkrgIlEY", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090453", "title": "Perception of Walking Self-body Avatar Enhances Virtual-walking Sensation", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090453/1jIxoojmMy4", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090577", "title": "Resizing of the peripersonal space for the seated for different step frequencies of vibrations at the soles", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090577/1jIxp3AAdhK", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2020/8508/0/850800a639", "title": "Visual-Auditory Redirection: Multimodal Integration of Incongruent Visual and Auditory Cues for Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/ismar/2020/850800a639/1pysvxeFG4E", "parentPublication": { "id": "proceedings/ismar/2020/8508/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2021/4057/0/405700a607", "title": "Virtual Walking Generator from Omnidirectional Video with Ground-dependent Foot Vibrations", "doi": null, "abstractUrl": "/proceedings-article/vrw/2021/405700a607/1tnWZe0CPwA", "parentPublication": { "id": "proceedings/vrw/2021/4057/0", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNANkoa6", "title": "2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops", "acronym": "cvprw", "groupId": "1001809", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNqJHFuT", "doi": "10.1109/CVPRW.2012.6238904", "title": "The measurement of eyestrain caused from diverse binocular disparities, viewing time and display sizes in watching stereoscopic 3D content", "normalizedTitle": "The measurement of eyestrain caused from diverse binocular disparities, viewing time and display sizes in watching stereoscopic 3D content", "abstract": "The measurement of eyestrain caused by watching stereoscopic 3D content is very important in producing 3D movies and 3D TV content that may be comfortably watched by the viewer. Some research has been done regarding comparative measurements of eyestrain between 2D and 3D video. However, what has not been explored sufficiently is the eyestrain caused by variations in binocular disparity, viewing time and display size in watching stereoscopic 3D video. In this paper, we quantified viewer discomfort by measuring eye blinking rates using an eye tracker system, and then estimated the user's eyestrain by integrating the eye blinking rates with user's subjective test responses when watching stereoscopic 3D content in relation to binocular disparity, viewing time and display size. As might have been expected, the experimental results show that eyestrain increases as the binocular disparity and viewing time increase. We also find that the viewer's eyestrain caused from watching stereoscopic 3D content on small displays is higher than with larger displays.", "abstracts": [ { "abstractType": "Regular", "content": "The measurement of eyestrain caused by watching stereoscopic 3D content is very important in producing 3D movies and 3D TV content that may be comfortably watched by the viewer. Some research has been done regarding comparative measurements of eyestrain between 2D and 3D video. However, what has not been explored sufficiently is the eyestrain caused by variations in binocular disparity, viewing time and display size in watching stereoscopic 3D video. In this paper, we quantified viewer discomfort by measuring eye blinking rates using an eye tracker system, and then estimated the user's eyestrain by integrating the eye blinking rates with user's subjective test responses when watching stereoscopic 3D content in relation to binocular disparity, viewing time and display size. As might have been expected, the experimental results show that eyestrain increases as the binocular disparity and viewing time increase. We also find that the viewer's eyestrain caused from watching stereoscopic 3D content on small displays is higher than with larger displays.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The measurement of eyestrain caused by watching stereoscopic 3D content is very important in producing 3D movies and 3D TV content that may be comfortably watched by the viewer. Some research has been done regarding comparative measurements of eyestrain between 2D and 3D video. However, what has not been explored sufficiently is the eyestrain caused by variations in binocular disparity, viewing time and display size in watching stereoscopic 3D video. In this paper, we quantified viewer discomfort by measuring eye blinking rates using an eye tracker system, and then estimated the user's eyestrain by integrating the eye blinking rates with user's subjective test responses when watching stereoscopic 3D content in relation to binocular disparity, viewing time and display size. As might have been expected, the experimental results show that eyestrain increases as the binocular disparity and viewing time increase. We also find that the viewer's eyestrain caused from watching stereoscopic 3D content on small displays is higher than with larger displays.", "fno": "06238904", "keywords": [ "Stereo Image Processing", "Video Signal Processing", "Eyestrain Measurement", "Diverse Binocular Disparities", "Display Sizes", "Viewing Time", "Stereoscopic 3 D Content", "3 D Movies", "3 D TV Content", "Comparative Measurements", "Stereoscopic 3 D Video", "Quantified Viewer", "Eye Blinking Rates", "Eye Tracker System", "Stereo Image Processing", "Three Dimensional Displays", "Visualization", "Strain", "Motion Pictures", "Fatigue", "Indexes" ], "authors": [ { "affiliation": "Dept. of Computer Engineering, The Catholic University of Korea #43-1 Yeokgok 2-dong, Wonmi-Gu, Bucheon, Gyeonggi-do, Korea", "fullName": "Sang-Hyun Cho", "givenName": "Sang-Hyun", "surname": "Cho", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept. of Digital Media, The Catholic University of Korea #43-1 Yeokgok 2-dong, Wonmi-Gu, Bucheon, Gyeonggi-do, Korea", "fullName": "Hang-Bong Kang", "givenName": "Hang-Bong", "surname": "Kang", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvprw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-06-01T00:00:00", "pubType": "proceedings", "pages": "23-28", "year": "2012", "issn": "2160-7508", "isbn": "978-1-4673-1611-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06238903", "articleId": "12OmNyRxFte", "__typename": "AdjacentArticleType" }, "next": { "fno": "06238905", "articleId": "12OmNAXxWUh", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/bdva/2015/7343/0/07314285", "title": "A Simple Objective Method for Automatic Error Detection in Stereoscopic 3D Video", "doi": null, "abstractUrl": "/proceedings-article/bdva/2015/07314285/12OmNC3FG9L", "parentPublication": { "id": "proceedings/bdva/2015/7343/0", "title": "2015 Big Data Visual Analytics (BDVA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/avss/2013/0703/0/06636637", "title": "Binocular video object tracking with fast disparity estimation", "doi": null, "abstractUrl": "/proceedings-article/avss/2013/06636637/12OmNrGsDqx", "parentPublication": { "id": "proceedings/avss/2013/0703/0", "title": "2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2014/4761/0/06890127", "title": "Binocular mismatch induced by luminance discrepancies on stereoscopic images", "doi": null, "abstractUrl": "/proceedings-article/icme/2014/06890127/12OmNvqmUBM", "parentPublication": { "id": "proceedings/icme/2014/4761/0", "title": "2014 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ssiai/2014/4053/0/06806030", "title": "Prediction of visual discomfort in watching 3D video using multiple features", "doi": null, "abstractUrl": "/proceedings-article/ssiai/2014/06806030/12OmNx6g6cF", "parentPublication": { "id": "proceedings/ssiai/2014/4053/0", "title": "2014 IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icinis/2012/4855/0/4855a302", "title": "Characteristic Point Match Algorithm Based on the SURF in Binocular Stereo Vision", "doi": null, "abstractUrl": "/proceedings-article/icinis/2012/4855a302/12OmNyxFKkc", "parentPublication": { "id": "proceedings/icinis/2012/4855/0", "title": "Intelligent Networks and Intelligent Systems, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/1993/3870/0/00378182", "title": "Direct estimation of multiple disparities for transparent multiple surfaces in binocular stereo", "doi": null, "abstractUrl": "/proceedings-article/iccv/1993/00378182/12OmNzSyCj1", "parentPublication": { "id": "proceedings/iccv/1993/3870/0", "title": "1993 (4th) International Conference on Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2014/4717/0/06890547", "title": "Supporting binocular visual quality prediction using machine learning", "doi": null, "abstractUrl": "/proceedings-article/icmew/2014/06890547/12OmNzWOB9v", "parentPublication": { "id": "proceedings/icmew/2014/4717/0", "title": "2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/1991/08/i0761", "title": "Direct Recovery of Three-Dimensional Scene Geometry From Binocular Stereo Disparity", "doi": null, "abstractUrl": "/journal/tp/1991/08/i0761/13rRUxDqS50", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bigmm/2018/5321/0/08499087", "title": "A 3D Visual Comfort Metric Based on Binocular Asymmetry Factor", "doi": null, "abstractUrl": "/proceedings-article/bigmm/2018/08499087/17D45VObpOA", "parentPublication": { "id": "proceedings/bigmm/2018/5321/0", "title": "2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000g654", "title": "Stereoscopic Neural Style Transfer", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000g654/17D45WaTkkG", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "13bd1eJgoia", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "13bd1fZBGdu", "doi": "10.1109/VR.2018.8448288", "title": "Experiencing an Invisible World War I Battlefield Through Narrative-Driven Redirected Walking in Virtual Reality", "normalizedTitle": "Experiencing an Invisible World War I Battlefield Through Narrative-Driven Redirected Walking in Virtual Reality", "abstract": "Redirected walking techniques have the potential to provide natural locomotion while users experience large virtual environments. However, when using redirected walking in small physical workspaces, disruptive overt resets are often required. We describe the design of an educational virtual reality experience in which users physically walk through virtual tunnels representative of the World War I battle of Vauquois. Walking in only a 15- by 5-foot tracked space, users are redirected through subtle, narrative-driven resets to walk through a tunnel nearly 50 feet in length. This work contributes approaches and lessons that can be used to provide a seamless and natural virtual reality walking experience in highly constrained physical spaces.", "abstracts": [ { "abstractType": "Regular", "content": "Redirected walking techniques have the potential to provide natural locomotion while users experience large virtual environments. However, when using redirected walking in small physical workspaces, disruptive overt resets are often required. We describe the design of an educational virtual reality experience in which users physically walk through virtual tunnels representative of the World War I battle of Vauquois. Walking in only a 15- by 5-foot tracked space, users are redirected through subtle, narrative-driven resets to walk through a tunnel nearly 50 feet in length. This work contributes approaches and lessons that can be used to provide a seamless and natural virtual reality walking experience in highly constrained physical spaces.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Redirected walking techniques have the potential to provide natural locomotion while users experience large virtual environments. However, when using redirected walking in small physical workspaces, disruptive overt resets are often required. We describe the design of an educational virtual reality experience in which users physically walk through virtual tunnels representative of the World War I battle of Vauquois. Walking in only a 15- by 5-foot tracked space, users are redirected through subtle, narrative-driven resets to walk through a tunnel nearly 50 feet in length. This work contributes approaches and lessons that can be used to provide a seamless and natural virtual reality walking experience in highly constrained physical spaces.", "fno": "08448288", "keywords": [ "Human Factors", "Tunnels", "User Interfaces", "Virtual Reality", "Invisible World War I Battlefield", "Highly Constrained Physical Spaces", "Natural Virtual Reality Walking Experience", "Seamless Reality Walking Experience", "Narrative Driven Resets", "5 Foot Tracked Space", "Virtual Tunnels Representative", "Educational Virtual Reality Experience", "Disruptive Overt Resets", "Physical Workspaces", "Virtual Environments", "Natural Locomotion", "Redirected Walking Techniques", "Legged Locomotion", "Virtual Environments", "Three Dimensional Displays", "Visualization", "Foot", "History", "Redirected Walking", "Narrative", "Educational VR H 5 1 Information Interfaces And Presentation E G HCI Multimedia Information Systems Artificial", "Augmented And Virtual Realities" ], "authors": [ { "affiliation": "Virginia Tech, Center for Human-Computer Interaction, Blacksburg, VA, USA", "fullName": "Run Yu", "givenName": "Run", "surname": "Yu", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, Center for Human-Computer Interaction, Blacksburg, VA, USA", "fullName": "Zachary Duer", "givenName": "Zachary", "surname": "Duer", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, Center for Human-Computer Interaction, Blacksburg, VA, USA", "fullName": "Todd Ogle", "givenName": "Todd", "surname": "Ogle", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, Center for Human-Computer Interaction, Blacksburg, VA, USA", "fullName": "Doug A. Bowman", "givenName": "Doug A.", "surname": "Bowman", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, Center for Human-Computer Interaction, Blacksburg, VA, USA", "fullName": "Thomas Tucker", "givenName": "Thomas", "surname": "Tucker", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, Center for Human-Computer Interaction, Blacksburg, VA, USA", "fullName": "David Hicks", "givenName": "David", "surname": "Hicks", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, School of Visual Arts, Blacksburg, VA, USA", "fullName": "Dongsoo Choi", "givenName": "Dongsoo", "surname": "Choi", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, School of Visual Arts, Blacksburg, VA, USA", "fullName": "Zach Bush", "givenName": "Zach", "surname": "Bush", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, School of Visual Arts, Blacksburg, VA, USA", "fullName": "Huy Ngo", "givenName": "Huy", "surname": "Ngo", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, School of Visual Arts, Blacksburg, VA, USA", "fullName": "Phat Nguyen", "givenName": "Phat", "surname": "Nguyen", "__typename": "ArticleAuthorType" }, { "affiliation": "Virginia Tech, School of Visual Arts, Blacksburg, VA, USA", "fullName": "Xindi Liu", "givenName": "Xindi", "surname": "Liu", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-03-01T00:00:00", "pubType": "proceedings", "pages": "313-319", "year": "2018", "issn": null, "isbn": "978-1-5386-3365-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08446216", "articleId": "13bd1gJ1v0k", "__typename": "AdjacentArticleType" }, "next": { "fno": "08447559", "articleId": "13bd1fHrlS0", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2013/4795/0/06549395", "title": "Flexible and general redirected walking for head-mounted displays", "doi": null, "abstractUrl": "/proceedings-article/vr/2013/06549395/12OmNxFJXN3", "parentPublication": { "id": "proceedings/vr/2013/4795/0", "title": "2013 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446479", "title": "Adopting the Roll Manipulation for Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446479/13bd1eSlys4", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2014/04/ttg201404579", "title": "Performance of Redirected Walking Algorithms in a Constrained Virtual World", "doi": null, "abstractUrl": "/journal/tg/2014/04/ttg201404579/13rRUwjoNx4", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/04/07036075", "title": "Cognitive Resource Demands of Redirected Walking", "doi": null, "abstractUrl": "/journal/tg/2015/04/07036075/13rRUxcKzVm", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/05/09715721", "title": "Validating Simulation-Based Evaluation of Redirected Walking Systems", "doi": null, "abstractUrl": "/journal/tg/2022/05/09715721/1B4hxt06P9m", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2022/11/09881577", "title": "Making Resets away from Targets: POI aware Redirected Walking", "doi": null, "abstractUrl": "/journal/tg/2022/11/09881577/1Gv8Ze0xuJG", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09961901", "title": "Transferable Virtual-Physical Environmental Alignment with Redirected Walking", "doi": null, "abstractUrl": "/journal/tg/5555/01/09961901/1IxvZ4KZbri", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798121", "title": "Real-time Optimal Planning for Redirected Walking Using Deep Q-Learning", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798121/1cJ17Y60ruM", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090595", "title": "Reactive Alignment of Virtual and Physical Environments Using Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090595/1jIxm1j8B2w", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/11/09523832", "title": "Redirected Walking in Static and Dynamic Scenes Using Visibility Polygons", "doi": null, "abstractUrl": "/journal/tg/2021/11/09523832/1wpqjiNuSqY", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "13bd1eJgoia", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "13bd1gJ1v0k", "doi": "10.1109/VR.2018.8446216", "title": "I Can See on My Feet While Walking: Sensitivity to Translation Gains with Visible Feet", "normalizedTitle": "I Can See on My Feet While Walking: Sensitivity to Translation Gains with Visible Feet", "abstract": "Redirected walking allows users to explore immersive virtual environments by real walking even when the physical tracking space is limited. Redirected walking is usually implemented via translation gains, rotation gains, and curvature gains, while previous research was focused on identifying detection thresholds for such manipulations. To our knowledge, all previous experiments were conducted without a visual self-representation of the user in the virtual environment, in particular, without showing the user's feet. In this paper, we address the question if the virtual self-representation of the user's feet changes the detection thresholds for translation gains. Furthermore, we consider the influence of the holisticness of the visual stimulus, i. e., the type of virtual environment. Therefore, we conducted an experiment to identify detection thresholds for translation gains under three different conditions: (i) without visible virtual feet and (ii) with visible virtual feet both in a high fidelity visually rich virtual environment, and (iii) with visible virtual feet in a low cue virtual environment. The results revealed the range of detection thresholds for translations gains, which cannot be detected by the user when the feet are visible. Furthermore, the results show a significant difference between the two types of environment. Our findings suggest that the virtual environment is more important for manipulation detection than the visual self-representation of the user's feet.", "abstracts": [ { "abstractType": "Regular", "content": "Redirected walking allows users to explore immersive virtual environments by real walking even when the physical tracking space is limited. Redirected walking is usually implemented via translation gains, rotation gains, and curvature gains, while previous research was focused on identifying detection thresholds for such manipulations. To our knowledge, all previous experiments were conducted without a visual self-representation of the user in the virtual environment, in particular, without showing the user's feet. In this paper, we address the question if the virtual self-representation of the user's feet changes the detection thresholds for translation gains. Furthermore, we consider the influence of the holisticness of the visual stimulus, i. e., the type of virtual environment. Therefore, we conducted an experiment to identify detection thresholds for translation gains under three different conditions: (i) without visible virtual feet and (ii) with visible virtual feet both in a high fidelity visually rich virtual environment, and (iii) with visible virtual feet in a low cue virtual environment. The results revealed the range of detection thresholds for translations gains, which cannot be detected by the user when the feet are visible. Furthermore, the results show a significant difference between the two types of environment. Our findings suggest that the virtual environment is more important for manipulation detection than the visual self-representation of the user's feet.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Redirected walking allows users to explore immersive virtual environments by real walking even when the physical tracking space is limited. Redirected walking is usually implemented via translation gains, rotation gains, and curvature gains, while previous research was focused on identifying detection thresholds for such manipulations. To our knowledge, all previous experiments were conducted without a visual self-representation of the user in the virtual environment, in particular, without showing the user's feet. In this paper, we address the question if the virtual self-representation of the user's feet changes the detection thresholds for translation gains. Furthermore, we consider the influence of the holisticness of the visual stimulus, i. e., the type of virtual environment. Therefore, we conducted an experiment to identify detection thresholds for translation gains under three different conditions: (i) without visible virtual feet and (ii) with visible virtual feet both in a high fidelity visually rich virtual environment, and (iii) with visible virtual feet in a low cue virtual environment. The results revealed the range of detection thresholds for translations gains, which cannot be detected by the user when the feet are visible. Furthermore, the results show a significant difference between the two types of environment. Our findings suggest that the virtual environment is more important for manipulation detection than the visual self-representation of the user's feet.", "fno": "08446216", "keywords": [ "Human Computer Interaction", "Virtual Reality", "Redirected Walking", "Immersive Virtual Environments", "Rotation Gains", "Curvature Gains", "Visible Virtual Feet", "High Fidelity Visually Rich Virtual Environment", "Low Cue Virtual Environment", "Feet Virtual Self Representation", "Feet Visual Self Representation", "Translation Gains Sensitivity", "Legged Locomotion", "Visualization", "Foot", "Virtual Environments", "Tracking", "Cameras", "Avatars", "Locomotion", "Redirected Walking", "Translation Gains H 5 1 Information Interfaces And Presentation Multimedia Information Systems X 2015 Artificial", "Augmented", "And Virtual Realities", "I 3 7 Computer Graphics Three Dimensional Graphics And Realism X 2015 Virtual Reality" ], "authors": [ { "affiliation": "University of Hamburg, Human-Computer Interaction", "fullName": "Lucie Kruse", "givenName": "Lucie", "surname": "Kruse", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Hamburg, Human-Computer Interaction", "fullName": "Eike Langbehn", "givenName": "Eike", "surname": "Langbehn", "__typename": "ArticleAuthorType" }, { "affiliation": "Human-Comput. Interaction, Univ. of Hamburg, Hamburg, Germany", "fullName": "Frank Steinicke", "givenName": "Frank", "surname": "Steinicke", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-03-01T00:00:00", "pubType": "proceedings", "pages": "305-312", "year": "2018", "issn": null, "isbn": "978-1-5386-3365-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08448290", "articleId": "13bd1AIBM21", "__typename": "AdjacentArticleType" }, "next": { "fno": "08448288", "articleId": "13bd1fZBGdu", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2015/1727/0/07223423", "title": "Tracking human locomotion by relative positional feet tracking", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223423/12OmNAZOJVa", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2014/3624/0/06798845", "title": "Feet movement in desktop 3D interaction", "doi": null, "abstractUrl": "/proceedings-article/3dui/2014/06798845/12OmNqC2uZJ", "parentPublication": { "id": "proceedings/3dui/2014/3624/0", "title": "2014 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2016/0836/0/07504743", "title": "Estimation of detection thresholds for audiovisual rotation gains", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504743/12OmNzmcm0b", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446479", "title": "Adopting the Roll Manipulation for Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446479/13bd1eSlys4", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446062", "title": "Biomechanical Parameters Under Curvature Gains and Bending Gains in Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446062/13bd1fKQxrR", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2018/7459/0/745900a115", "title": "Rethinking Redirected Walking: On the Use of Curvature Gains Beyond Perceptual Limitations and Revisiting Bending Gains", "doi": null, "abstractUrl": "/proceedings-article/ismar/2018/745900a115/17D45WK5AlG", "parentPublication": { "id": "proceedings/ismar/2018/7459/0", "title": "2018 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2022/8402/0/840200a830", "title": "Redirected Walking in 360° Video: Effect of Environment Size on Detection Thresholds for Translation and Rotation Gains", "doi": null, "abstractUrl": "/proceedings-article/vrw/2022/840200a830/1CJd1TReEYo", "parentPublication": { "id": "proceedings/vrw/2022/8402/0", "title": "2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798345", "title": "Investigation of Visual Self-Representation for a Walking-in-Place Navigation System in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798345/1cJ1hpkUgHS", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090671", "title": "The Influence of Full-Body Representation on Translation and Curvature Gain", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090671/1jIxqcIwi64", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2021/4057/0/405700a358", "title": "Revisiting Audiovisual Rotation Gains for Redirected Walking", "doi": null, "abstractUrl": "/proceedings-article/vrw/2021/405700a358/1tnXe22MFJm", "parentPublication": { "id": "proceedings/vrw/2021/4057/0", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "17D45VtKirz", "title": "2018 IEEE 4th VR Workshop on Sonic Interactions for Virtual Environments (SIVE)", "acronym": "sive", "groupId": "1805064", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "17D45XoXP3w", "doi": "10.1109/SIVE.2018.8577177", "title": "Influence of hearing your steps and environmental sounds in VR while walking", "normalizedTitle": "Influence of hearing your steps and environmental sounds in VR while walking", "abstract": "Presence, the feeling of `being there' in a virtual environment (VR), is seen as a basic requirement for VR environments.In a controlled laboratory experiment, the effect of auditory components of VR on presence was studied. Participants wore a head-mounted display (Oculus Rift DK2) and noise-cancelling headphones (Bose Quiet Comfort 25) while walking on a treadmill for 5 minutes per experimental condition. They were visually presented with a rural virtual reality scenery. The audio presentation was varied in a 2×2 within-subjects design with a \"Soundscape\" (birds, wind, church bells) and/or \"Footsteps\" (played back when the sensors detected a step), either being presented or not. After each condition, the subjects filled out the IPQ Presence Questionnaire.Data analysis was conducted on 40 participants who completed each of the four sound conditions in random order. The results of the statistical analysis show that (a) a single-item presence scale (percentage of `perfect' presence), (b) the IPQ presence scale G1, and (c) the IPQ realism scale all detected significant effects of playing back footstep sounds, and even larger effects due to providing a soundscape in VR. By contrast, IPQ spatial presence was not increased and IPQ involvement was only increased by providing the environmental soundscape.To sum up, the self-generated sounds had significant effects on the feeling of \"being there\" and on perceived realism, while the effect of the soundscape tended to be even more powerful.", "abstracts": [ { "abstractType": "Regular", "content": "Presence, the feeling of `being there' in a virtual environment (VR), is seen as a basic requirement for VR environments.In a controlled laboratory experiment, the effect of auditory components of VR on presence was studied. Participants wore a head-mounted display (Oculus Rift DK2) and noise-cancelling headphones (Bose Quiet Comfort 25) while walking on a treadmill for 5 minutes per experimental condition. They were visually presented with a rural virtual reality scenery. The audio presentation was varied in a 2×2 within-subjects design with a \"Soundscape\" (birds, wind, church bells) and/or \"Footsteps\" (played back when the sensors detected a step), either being presented or not. After each condition, the subjects filled out the IPQ Presence Questionnaire.Data analysis was conducted on 40 participants who completed each of the four sound conditions in random order. The results of the statistical analysis show that (a) a single-item presence scale (percentage of `perfect' presence), (b) the IPQ presence scale G1, and (c) the IPQ realism scale all detected significant effects of playing back footstep sounds, and even larger effects due to providing a soundscape in VR. By contrast, IPQ spatial presence was not increased and IPQ involvement was only increased by providing the environmental soundscape.To sum up, the self-generated sounds had significant effects on the feeling of \"being there\" and on perceived realism, while the effect of the soundscape tended to be even more powerful.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Presence, the feeling of `being there' in a virtual environment (VR), is seen as a basic requirement for VR environments.In a controlled laboratory experiment, the effect of auditory components of VR on presence was studied. Participants wore a head-mounted display (Oculus Rift DK2) and noise-cancelling headphones (Bose Quiet Comfort 25) while walking on a treadmill for 5 minutes per experimental condition. They were visually presented with a rural virtual reality scenery. The audio presentation was varied in a 2×2 within-subjects design with a \"Soundscape\" (birds, wind, church bells) and/or \"Footsteps\" (played back when the sensors detected a step), either being presented or not. After each condition, the subjects filled out the IPQ Presence Questionnaire.Data analysis was conducted on 40 participants who completed each of the four sound conditions in random order. The results of the statistical analysis show that (a) a single-item presence scale (percentage of `perfect' presence), (b) the IPQ presence scale G1, and (c) the IPQ realism scale all detected significant effects of playing back footstep sounds, and even larger effects due to providing a soundscape in VR. By contrast, IPQ spatial presence was not increased and IPQ involvement was only increased by providing the environmental soundscape.To sum up, the self-generated sounds had significant effects on the feeling of \"being there\" and on perceived realism, while the effect of the soundscape tended to be even more powerful.", "fno": "08577177", "keywords": [ "Data Analysis", "Headphones", "Helmet Mounted Displays", "Statistical Analysis", "Virtual Reality", "Data Analysis", "G 1 IPQ Presence Scale", "Statistical Analysis", "Noise Cancelling Headphones", "IPQ Spatial Presence", "Single Item Presence Scale", "Audio Presentation", "Rural Virtual Reality Scenery", "Bose Quiet Comfort 25", "Oculus Rift DK 2", "Head Mounted Display", "Auditory Components", "VR Environments", "Virtual Environment", "Environmental Sounds", "Legged Locomotion", "Virtual Environments", "Head Mounted Displays", "Headphones", "Foot", "Standards", "Virtual Reality", "Presence", "Soundscape", "Footsteps", "IPQ", "Head Mounted Display" ], "authors": [ { "affiliation": "Technische Universität Darmstadt", "fullName": "Angelika C. Kern", "givenName": "Angelika C.", "surname": "Kern", "__typename": "ArticleAuthorType" }, { "affiliation": "Technische Universität Darmstadt", "fullName": "Wolfgang Ellermeier", "givenName": "Wolfgang", "surname": "Ellermeier", "__typename": "ArticleAuthorType" } ], "idPrefix": "sive", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-03-01T00:00:00", "pubType": "proceedings", "pages": "1-4", "year": "2018", "issn": null, "isbn": "978-1-5386-5713-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08577076", "articleId": "17D45VsBU4f", "__typename": "AdjacentArticleType" }, "next": { "fno": "08577195", "articleId": "17D45XeKgwR", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "trans/tg/2018/04/08267106", "title": "Force Rendering and its Evaluation of a Friction-Based Walking Sensation Display for a Seated User", "doi": null, "abstractUrl": "/journal/tg/2018/04/08267106/13rRUwIF6dW", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/04/08260962", "title": "Ascending and Descending in Virtual Reality: Simple and Safe System Using Passive Haptics", "doi": null, "abstractUrl": "/journal/tg/2018/04/08260962/13rRUwjGoLM", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2018/7592/0/08699289", "title": "Walking-in-Place for VR Navigation Independent of Gaze Direction Using a Waist-Worn Inertial Measurement Unit", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2018/08699289/19F1PlWtKJa", "parentPublication": { "id": "proceedings/ismar-adjunct/2018/7592/0", "title": "2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09881908", "title": "PropelWalker: A Leg-Based Wearable System With Propeller-Based Force Feedback for Walking in Fluids in VR", "doi": null, "abstractUrl": "/journal/tg/5555/01/09881908/1Gv909WpCG4", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2023/4815/0/481500a094", "title": "An EEG-based Experiment on VR Sickness and Postural Instability While Walking in Virtual Environments", "doi": null, "abstractUrl": "/proceedings-article/vr/2023/481500a094/1MNgWtYsR5S", "parentPublication": { "id": "proceedings/vr/2023/4815/0", "title": "2023 IEEE Conference Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/01/08762207", "title": "Locomotion in Place in Virtual Reality: A Comparative Evaluation of Joystick, Teleport, and Leaning", "doi": null, "abstractUrl": "/journal/tg/2021/01/08762207/1bIeI0S82Aw", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08797999", "title": "VR system to simulate tightrope walking with a standalone VR headset and slack rails", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08797999/1cJ0Nqr10CA", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2020/8508/0/850800a608", "title": "Walking and Teleportation in Wide-area Virtual Reality Experiences", "doi": null, "abstractUrl": "/proceedings-article/ismar/2020/850800a608/1pysv8bIfrG", "parentPublication": { "id": "proceedings/ismar/2020/8508/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2020/8508/0/850800a649", "title": "Comparing World and Screen Coordinate Systems in Optical See-Through Head-Mounted Displays for Text Readability while Walking", "doi": null, "abstractUrl": "/proceedings-article/ismar/2020/850800a649/1pysvKFdazS", "parentPublication": { "id": "proceedings/ismar/2020/8508/0", "title": "2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/11/09523894", "title": "Head-Mounted Display with Increased Downward Field of View Improves Presence and Sense of Self-Location", "doi": null, "abstractUrl": "/journal/tg/2021/11/09523894/1wpqkPb7CSY", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "17D45VtKipM", "title": "2017 International Conference on Culture and Computing (Culture and Computing)", "acronym": "culture-and-computing", "groupId": "1800597", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "17D45XtvpdY", "doi": "10.1109/Culture.and.Computing.2017.25", "title": "Walk through a Museum with Binocular Stereo Effect and Spherical Panorama Views", "normalizedTitle": "Walk through a Museum with Binocular Stereo Effect and Spherical Panorama Views", "abstract": "Photo based spherical panoramas were usually used to represent scenes in art museum, but traditional panorama only allows looking around the scene from a fixed point and it could not provide binocular stereo effect that is very necessary to build up a more immersive experience for art appreciation. Modeling based virtual tour scene could provide walking through ability and binocular stereo effect but it is almost impossible to model an art museum as just it likes. In this paper, the authors designed a special method for spherical panorama recording, based on which kind of virtual tour system was developed to provide walking through ability and binocular stereo effect to users with head mounted display. This system combines the advantages of photo realistic, walking through and binocular stereo effect.", "abstracts": [ { "abstractType": "Regular", "content": "Photo based spherical panoramas were usually used to represent scenes in art museum, but traditional panorama only allows looking around the scene from a fixed point and it could not provide binocular stereo effect that is very necessary to build up a more immersive experience for art appreciation. Modeling based virtual tour scene could provide walking through ability and binocular stereo effect but it is almost impossible to model an art museum as just it likes. In this paper, the authors designed a special method for spherical panorama recording, based on which kind of virtual tour system was developed to provide walking through ability and binocular stereo effect to users with head mounted display. This system combines the advantages of photo realistic, walking through and binocular stereo effect.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Photo based spherical panoramas were usually used to represent scenes in art museum, but traditional panorama only allows looking around the scene from a fixed point and it could not provide binocular stereo effect that is very necessary to build up a more immersive experience for art appreciation. Modeling based virtual tour scene could provide walking through ability and binocular stereo effect but it is almost impossible to model an art museum as just it likes. In this paper, the authors designed a special method for spherical panorama recording, based on which kind of virtual tour system was developed to provide walking through ability and binocular stereo effect to users with head mounted display. This system combines the advantages of photo realistic, walking through and binocular stereo effect.", "fno": "08227335", "keywords": [ "Art", "Helmet Mounted Displays", "Museums", "Stereo Image Processing", "Virtual Reality", "Binocular Stereo Effect", "Art Museum", "Spherical Panorama Recording", "Cameras", "Lenses", "Legged Locomotion", "Three Dimensional Displays", "Rails", "Rendering Computer Graphics", "Art", "Art Museum", "Spherical Panorama", "Walking Through", "Binocular Stereo Effect" ], "authors": [ { "affiliation": null, "fullName": "YanXiang Zhang", "givenName": "YanXiang", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "ZiQiang Zhu", "givenName": "ZiQiang", "surname": "Zhu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "PengFei Ma", "givenName": "PengFei", "surname": "Ma", "__typename": "ArticleAuthorType" } ], "idPrefix": "culture-and-computing", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-09-01T00:00:00", "pubType": "proceedings", "pages": "20-23", "year": "2017", "issn": null, "isbn": "978-1-5386-1135-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08227334", "articleId": "17D45XERmmb", "__typename": "AdjacentArticleType" }, "next": { "fno": "08227336", "articleId": "17D45VsBTYR", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2015/1727/0/07223431", "title": "Walking recording and experience system by Visual Psychophysics Lab", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223431/12OmNB1NVNQ", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2011/0063/0/06130256", "title": "3D environment measurement using binocular stereo and motion stereo by mobile robot with omnidirectional stereo camera", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2011/06130256/12OmNBl6EHn", "parentPublication": { "id": "proceedings/iccvw/2011/0063/0", "title": "2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2002/1695/3/169530635", "title": "The Design of a Stereo Panorama Camera for Scenes of Dynamic Range", "doi": null, "abstractUrl": "/proceedings-article/icpr/2002/169530635/12OmNvlg8kj", "parentPublication": { "id": "proceedings/icpr/2002/1695/3", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2017/6647/0/07892287", "title": "Evaluation of airflow effect on a VR walk", "doi": null, "abstractUrl": "/proceedings-article/vr/2017/07892287/12OmNwtEEvF", "parentPublication": { "id": "proceedings/vr/2017/6647/0", "title": "2017 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscc/2016/0679/0/07543892", "title": "Scenic Athens: A personalized scenic route planner for tourists", "doi": null, "abstractUrl": "/proceedings-article/iscc/2016/07543892/12OmNwwMf25", "parentPublication": { "id": "proceedings/iscc/2016/0679/0", "title": "2016 IEEE Symposium on Computers and Communication (ISCC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/culture-computing/2015/8232/0/8232a199", "title": "Virtual Show, Go In!: Walk-Through System and VR Goggles of a Temple for Museum Exhibits", "doi": null, "abstractUrl": "/proceedings-article/culture-computing/2015/8232a199/12OmNzkMlVq", "parentPublication": { "id": "proceedings/culture-computing/2015/8232/0", "title": "2015 International Conference on Culture and Computing (Culture Computing)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446426", "title": "Walk-Centric User Interfaces", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446426/13bd1fHrlRZ", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmtma/2022/9978/0/997800a718", "title": "A probabilistic model for rail transit passenger flow distribution using AFC data", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2022/997800a718/1ByeQaliKnS", "parentPublication": { "id": "proceedings/icmtma/2022/9978/0", "title": "2022 14th International Conference on Measuring Technology and Mechatronics Automation (ICMTMA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08797999", "title": "VR system to simulate tightrope walking with a standalone VR headset and slack rails", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08797999/1cJ0Nqr10CA", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08967011", "title": "The Role of Binocular Vision in Avoiding Virtual Obstacles While Walking", "doi": null, "abstractUrl": "/journal/tg/2021/07/08967011/1gPjyDVBxF6", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1cI6akLvAuQ", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1cJ0WSuJ27e", "doi": "10.1109/VR.2019.8797751", "title": "Improving Walking in Place Methods with Individualization and Deep Networks", "normalizedTitle": "Improving Walking in Place Methods with Individualization and Deep Networks", "abstract": "Walking in place is a standard method for moving through large virtual environments when physical space or positional tracking is limited. This technique has become increasingly prominent with the advent of mobile virtual reality in which external tracking may not be present. In this paper, we revisit walking in place algorithms to address some of their technical challenges. Namely, our solutions attend to improving starting, stopping, and speed control for individual users. From a hand-tuned threshold based algorithm, we provide a new, fast method for individualizing the walking in place algorithm based on biomechanic measures of step rate. In addition, we introduce a new walking in place model based on a convolutional neural network trained to differentiate walking and standing. Over two experiments we assess these methods against a traditional threshold based algorithm on two mobile virtual reality platforms. The assessments are based on controllability, scale, and presence. Our results suggest that an adequately trained convolutional neural network can be an effective way of implementing walking in place.", "abstracts": [ { "abstractType": "Regular", "content": "Walking in place is a standard method for moving through large virtual environments when physical space or positional tracking is limited. This technique has become increasingly prominent with the advent of mobile virtual reality in which external tracking may not be present. In this paper, we revisit walking in place algorithms to address some of their technical challenges. Namely, our solutions attend to improving starting, stopping, and speed control for individual users. From a hand-tuned threshold based algorithm, we provide a new, fast method for individualizing the walking in place algorithm based on biomechanic measures of step rate. In addition, we introduce a new walking in place model based on a convolutional neural network trained to differentiate walking and standing. Over two experiments we assess these methods against a traditional threshold based algorithm on two mobile virtual reality platforms. The assessments are based on controllability, scale, and presence. Our results suggest that an adequately trained convolutional neural network can be an effective way of implementing walking in place.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Walking in place is a standard method for moving through large virtual environments when physical space or positional tracking is limited. This technique has become increasingly prominent with the advent of mobile virtual reality in which external tracking may not be present. In this paper, we revisit walking in place algorithms to address some of their technical challenges. Namely, our solutions attend to improving starting, stopping, and speed control for individual users. From a hand-tuned threshold based algorithm, we provide a new, fast method for individualizing the walking in place algorithm based on biomechanic measures of step rate. In addition, we introduce a new walking in place model based on a convolutional neural network trained to differentiate walking and standing. Over two experiments we assess these methods against a traditional threshold based algorithm on two mobile virtual reality platforms. The assessments are based on controllability, scale, and presence. Our results suggest that an adequately trained convolutional neural network can be an effective way of implementing walking in place.", "fno": "08797751", "keywords": [ "Biomechanics", "Convolutional Neural Nets", "Learning Artificial Intelligence", "Virtual Reality", "Virtual Environments", "Physical Space", "Positional Tracking", "External Tracking", "Speed Control", "Hand Tuned Threshold Based Algorithm", "Place Algorithm", "Biomechanic Measures", "Step Rate", "Place Model", "Differentiate Walking", "Mobile Virtual Reality Platforms", "Walking Improvement", "Trained Convolutional Neural Network", "Legged Locomotion", "Virtual Environments", "Tracking", "Acceleration", "Gears", "Neural Networks", "Magnetic Heads", "Virtual Environments", "Locomotion", "Walking In Place", "Convolutional Neural Network", "Perception", "I 3 7 Computer Graphics Three Dimensional Graphics And Realism X 2014 Virtual Reality", "J 4 Computer Applications Social And Behavioral Sciences X 2014 Psychology" ], "authors": [ { "affiliation": "University of Southern, California, USA", "fullName": "Sara Hanson", "givenName": "Sara", "surname": "Hanson", "__typename": "ArticleAuthorType" }, { "affiliation": "Vanderbilt University, USA", "fullName": "Richard A. Paris", "givenName": "Richard A.", "surname": "Paris", "__typename": "ArticleAuthorType" }, { "affiliation": "Vanderbilt University, USA", "fullName": "Haley A. Adams", "givenName": "Haley A.", "surname": "Adams", "__typename": "ArticleAuthorType" }, { "affiliation": "Vanderbilt University, USA", "fullName": "Bobby Bodenheimer", "givenName": "Bobby", "surname": "Bodenheimer", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-03-01T00:00:00", "pubType": "proceedings", "pages": "367-376", "year": "2019", "issn": null, "isbn": "978-1-7281-1377-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08798074", "articleId": "1cJ0OPBhW4U", "__typename": "AdjacentArticleType" }, "next": { "fno": "08797994", "articleId": "1cJ19tjOG2s", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2010/6237/0/05444812", "title": "GUD WIP: Gait-Understanding-Driven Walking-In-Place", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444812/12OmNAle6ku", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2014/3624/0/06798850", "title": "A comparison of different methods for reducing the unintended positional drift accompanying walking-in-place locomotion", "doi": null, "abstractUrl": "/proceedings-article/3dui/2014/06798850/12OmNvCzFbu", "parentPublication": { "id": "proceedings/3dui/2014/3624/0", "title": "2014 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2016/0836/0/07504754", "title": "Detecting movement patterns from inertial data of a mobile head-mounted-display for navigation via walking-in-place", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504754/12OmNxWLTrY", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2008/2047/0/04476598", "title": "LLCM-WIP: Low-Latency, Continuous-Motion Walking-in-Place", "doi": null, "abstractUrl": "/proceedings-article/3dui/2008/04476598/12OmNyQYtvN", "parentPublication": { "id": "proceedings/3dui/2008/2047/0", "title": "2008 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2014/04/ttg201404569", "title": "Establishing the Range of Perceptually Natural Visual Walking Speeds for Virtual Walking-In-Place Locomotion", "doi": null, "abstractUrl": "/journal/tg/2014/04/ttg201404569/13rRUxAASTb", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09978713", "title": "Revisiting Walking-in-Place by Introducing Step-Height Control, Elastic Input, and Pseudo-Haptic Feedback", "doi": null, "abstractUrl": "/journal/tg/5555/01/09978713/1IXUnnVaWoE", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2023/05/10049680", "title": "Assisted walking-in-place: Introducing assisted motion to walking-by-cycling in embodied virtual reality", "doi": null, "abstractUrl": "/journal/tg/2023/05/10049680/1KYolEFtr6U", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798345", "title": "Investigation of Visual Self-Representation for a Walking-in-Place Navigation System in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798345/1cJ1hpkUgHS", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08967011", "title": "The Role of Binocular Vision in Avoiding Virtual Obstacles While Walking", "doi": null, "abstractUrl": "/journal/tg/2021/07/08967011/1gPjyDVBxF6", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2020/5608/0/09089561", "title": "Real Walking in Place: HEX-CORE-PROTOTYPE Omnidirectional Treadmill", "doi": null, "abstractUrl": "/proceedings-article/vr/2020/09089561/1jIxfncHjNe", "parentPublication": { "id": "proceedings/vr/2020/5608/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1cI6akLvAuQ", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "acronym": "vr", "groupId": "1000791", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1cJ1hpkUgHS", "doi": "10.1109/VR.2019.8798345", "title": "Investigation of Visual Self-Representation for a Walking-in-Place Navigation System in Virtual Reality", "normalizedTitle": "Investigation of Visual Self-Representation for a Walking-in-Place Navigation System in Virtual Reality", "abstract": "Walking-in-place (WIP) is one of the techniques for navigation in virtual reality (VR), and it can be configured in a limited space with a simple algorithm. Although WIP systems provide a sense of movement, it is important to deliver immersive VR experiences by providing information as similar as possible to walking in the real world. There have been many studies on the WIP technology, but it has rarely been done on visual self-representation of WIP in the virtual environment (VE). In this paper, we describe our investigation of virtual self-representation for application to a WIP navigation system using a HMD and full body motion capture system. Our system is designed to move in the pelvis direction by calculating the inertial sensor data, and a virtual body that is linked to the user's movement is seen from the first-person perspective (1PP) in two ways: (i) full body, and (ii) full body with natural walking. In (ii), when a step is detected, the motion of the lower part of the avatar is manipulated as if the user is performing real walking. We discuss the possibility of visual self-representation for the WIP system.", "abstracts": [ { "abstractType": "Regular", "content": "Walking-in-place (WIP) is one of the techniques for navigation in virtual reality (VR), and it can be configured in a limited space with a simple algorithm. Although WIP systems provide a sense of movement, it is important to deliver immersive VR experiences by providing information as similar as possible to walking in the real world. There have been many studies on the WIP technology, but it has rarely been done on visual self-representation of WIP in the virtual environment (VE). In this paper, we describe our investigation of virtual self-representation for application to a WIP navigation system using a HMD and full body motion capture system. Our system is designed to move in the pelvis direction by calculating the inertial sensor data, and a virtual body that is linked to the user's movement is seen from the first-person perspective (1PP) in two ways: (i) full body, and (ii) full body with natural walking. In (ii), when a step is detected, the motion of the lower part of the avatar is manipulated as if the user is performing real walking. We discuss the possibility of visual self-representation for the WIP system.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Walking-in-place (WIP) is one of the techniques for navigation in virtual reality (VR), and it can be configured in a limited space with a simple algorithm. Although WIP systems provide a sense of movement, it is important to deliver immersive VR experiences by providing information as similar as possible to walking in the real world. There have been many studies on the WIP technology, but it has rarely been done on visual self-representation of WIP in the virtual environment (VE). In this paper, we describe our investigation of virtual self-representation for application to a WIP navigation system using a HMD and full body motion capture system. Our system is designed to move in the pelvis direction by calculating the inertial sensor data, and a virtual body that is linked to the user's movement is seen from the first-person perspective (1PP) in two ways: (i) full body, and (ii) full body with natural walking. In (ii), when a step is detected, the motion of the lower part of the avatar is manipulated as if the user is performing real walking. We discuss the possibility of visual self-representation for the WIP system.", "fno": "08798345", "keywords": [ "Avatars", "Helmet Mounted Displays", "Visual Self Representation", "Walking In Place Navigation System", "Virtual Reality", "Immersive VR Experiences", "Virtual Self Representation", "WIP Navigation System", "Virtual Body", "Natural Walking", "HMD", "Full Body Motion Capture System", "Pelvis Direction", "Inertial Sensor Data", "First Person Perspective", "Avatar", "Foot", "Legged Locomotion", "Avatars", "Visualization", "Navigation", "Virtual Environments", "Computing Methodologies", "Computer Graphics", "Graphics Systems And Interfaces", "Virtual Reality", "Human Centered Computing", "Human Computer Interaction HCI", "Interaction Techniques" ], "authors": [ { "affiliation": "Electronics and Telecommunications Research Institute", "fullName": "Chanho Park", "givenName": "Chanho", "surname": "Park", "__typename": "ArticleAuthorType" }, { "affiliation": "Electronics and Telecommunications Research Institute", "fullName": "Kyungho Jang", "givenName": "Kyungho", "surname": "Jang", "__typename": "ArticleAuthorType" } ], "idPrefix": "vr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-03-01T00:00:00", "pubType": "proceedings", "pages": "1114-1115", "year": "2019", "issn": null, "isbn": "978-1-7281-1377-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08798235", "articleId": "1cJ0IDINHJm", "__typename": "AdjacentArticleType" }, "next": { "fno": "08798168", "articleId": "1cJ1h486id2", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2010/6237/0/05444812", "title": "GUD WIP: Gait-Understanding-Driven Walking-In-Place", "doi": null, "abstractUrl": "/proceedings-article/vr/2010/05444812/12OmNAle6ku", "parentPublication": { "id": "proceedings/vr/2010/6237/0", "title": "2010 IEEE Virtual Reality Conference (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2013/6097/0/06550193", "title": "Tapping-In-Place: Increasing the naturalness of immersive walking-in-place locomotion through novel gestural input", "doi": null, "abstractUrl": "/proceedings-article/3dui/2013/06550193/12OmNAnMuyq", "parentPublication": { "id": "proceedings/3dui/2013/6097/0", "title": "2013 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2008/2047/0/04476598", "title": "LLCM-WIP: Low-Latency, Continuous-Motion Walking-in-Place", "doi": null, "abstractUrl": "/proceedings-article/3dui/2008/04476598/12OmNyQYtvN", "parentPublication": { "id": "proceedings/3dui/2008/2047/0", "title": "2008 IEEE Symposium on 3D User Interfaces", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446216", "title": "I Can See on My Feet While Walking: Sensitivity to Translation Gains with Visible Feet", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446216/13bd1gJ1v0k", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2014/04/ttg201404569", "title": "Establishing the Range of Perceptually Natural Visual Walking Speeds for Virtual Walking-In-Place Locomotion", "doi": null, "abstractUrl": "/journal/tg/2014/04/ttg201404569/13rRUxAASTb", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2018/7592/0/08699289", "title": "Walking-in-Place for VR Navigation Independent of Gaze Direction Using a Waist-Worn Inertial Measurement Unit", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2018/08699289/19F1PlWtKJa", "parentPublication": { "id": "proceedings/ismar-adjunct/2018/7592/0", "title": "2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09978713", "title": "Revisiting Walking-in-Place by Introducing Step-Height Control, Elastic Input, and Pseudo-Haptic Feedback", "doi": null, "abstractUrl": "/journal/tg/5555/01/09978713/1IXUnnVaWoE", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2023/05/10049680", "title": "Assisted walking-in-place: Introducing assisted motion to walking-by-cycling in embodied virtual reality", "doi": null, "abstractUrl": "/journal/tg/2023/05/10049680/1KYolEFtr6U", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090453", "title": "Perception of Walking Self-body Avatar Enhances Virtual-walking Sensation", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090453/1jIxoojmMy4", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090671", "title": "The Influence of Full-Body Representation on Translation and Curvature Gain", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090671/1jIxqcIwi64", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1jIxhEnA8IE", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "acronym": "vrw", "groupId": "1836626", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1jIxoojmMy4", "doi": "10.1109/VRW50115.2020.00217", "title": "Perception of Walking Self-body Avatar Enhances Virtual-walking Sensation", "normalizedTitle": "Perception of Walking Self-body Avatar Enhances Virtual-walking Sensation", "abstract": "Various virtual walking systems have been developed using treadmill or leg-support devices to move users’ legs. We proposed and evaluated a virtual walking system for sitting observers using only passive sensations such as optic flow and foot vibrations. In this study, we examined whether the virtual body representing observer’s walking could facilitate the sensations of virtual walking in a virtual environment. We found that the virtual body enhanced sensations of walking, leg action and presence except for self-motion. These results suggest that perception of walking self-body avatar would facilitate sensations of walking and presence independently from foot vibrations.", "abstracts": [ { "abstractType": "Regular", "content": "Various virtual walking systems have been developed using treadmill or leg-support devices to move users’ legs. We proposed and evaluated a virtual walking system for sitting observers using only passive sensations such as optic flow and foot vibrations. In this study, we examined whether the virtual body representing observer’s walking could facilitate the sensations of virtual walking in a virtual environment. We found that the virtual body enhanced sensations of walking, leg action and presence except for self-motion. These results suggest that perception of walking self-body avatar would facilitate sensations of walking and presence independently from foot vibrations.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Various virtual walking systems have been developed using treadmill or leg-support devices to move users’ legs. We proposed and evaluated a virtual walking system for sitting observers using only passive sensations such as optic flow and foot vibrations. In this study, we examined whether the virtual body representing observer’s walking could facilitate the sensations of virtual walking in a virtual environment. We found that the virtual body enhanced sensations of walking, leg action and presence except for self-motion. These results suggest that perception of walking self-body avatar would facilitate sensations of walking and presence independently from foot vibrations.", "fno": "09090453", "keywords": [ "Legged Locomotion", "Avatars", "Foot", "Vibrations", "Telepresence", "Optical Flow", "Virtual Environments", "Self Motion", "Walking", "Optic Flow", "Avatar", "Tactile Sensation" ], "authors": [ { "affiliation": "Toyohashi University of Technology", "fullName": "Yusuke Matsuda", "givenName": "Yusuke", "surname": "Matsuda", "__typename": "ArticleAuthorType" }, { "affiliation": "Toyohashi University of Technology", "fullName": "Junya Nakamura", "givenName": "Junya", "surname": "Nakamura", "__typename": "ArticleAuthorType" }, { "affiliation": "The University of Tokyo", "fullName": "Tomohiro Amemiya", "givenName": "Tomohiro", "surname": "Amemiya", "__typename": "ArticleAuthorType" }, { "affiliation": "Tokyo Metropolitan University", "fullName": "Yasushi Ikei", "givenName": "Yasushi", "surname": "Ikei", "__typename": "ArticleAuthorType" }, { "affiliation": "Toyohashi University of Technology", "fullName": "Michiteru Kitazaki", "givenName": "Michiteru", "surname": "Kitazaki", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-03-01T00:00:00", "pubType": "proceedings", "pages": "732-733", "year": "2020", "issn": null, "isbn": "978-1-7281-6532-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09090643", "articleId": "1jIxuSITfqM", "__typename": "AdjacentArticleType" }, "next": { "fno": "09090414", "articleId": "1jIxvrGoeFq", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vr/2015/1727/0/07223431", "title": "Walking recording and experience system by Visual Psychophysics Lab", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223431/12OmNB1NVNQ", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2016/0842/0/07460066", "title": "Rhythmic vibrations to heels and forefeet to produce virtual walking", "doi": null, "abstractUrl": "/proceedings-article/3dui/2016/07460066/12OmNBQkwZJ", "parentPublication": { "id": "proceedings/3dui/2016/0842/0", "title": "2016 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dui/2012/1204/0/06184179", "title": "The King-Kong Effects: Improving sensation of walking in VR with visual and tactile vibrations at each step", "doi": null, "abstractUrl": "/proceedings-article/3dui/2012/06184179/12OmNwEJ0Lv", "parentPublication": { "id": "proceedings/3dui/2012/1204/0", "title": "2012 IEEE Symposium on 3D User Interfaces (3DUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2016/0836/0/07504715", "title": "Vestibulohaptic passive stimulation for a walking sensation", "doi": null, "abstractUrl": "/proceedings-article/vr/2016/07504715/12OmNxu6p8R", "parentPublication": { "id": "proceedings/vr/2016/0836/0", "title": "2016 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/med/2006/1/0/04124877", "title": "Observer-based control for absolute orientation estimation of a five-link walking biped robot", "doi": null, "abstractUrl": "/proceedings-article/med/2006/04124877/12OmNz5JBNP", "parentPublication": { "id": "proceedings/med/2006/1/0", "title": "Proceedings of the 14th Mediterranean Conference on Control and Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ta/2017/03/07450621", "title": "Emotion Rendering in Plantar Vibro-Tactile Simulations of Imagined Walking Styles", "doi": null, "abstractUrl": "/journal/ta/2017/03/07450621/13rRUwIF6cq", "parentPublication": { "id": "trans/ta", "title": "IEEE Transactions on Affective Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09911682", "title": "Effect of Vibrations on Impression of Walking and Embodiment With First- and Third-Person Avatar", "doi": null, "abstractUrl": "/journal/tg/5555/01/09911682/1HeiWQWKlTG", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798345", "title": "Investigation of Visual Self-Representation for a Walking-in-Place Navigation System in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798345/1cJ1hpkUgHS", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2020/6532/0/09090634", "title": "Rhythmic proprioceptive stimulation improves embodiment in a walking avatar when added to visual stimulation", "doi": null, "abstractUrl": "/proceedings-article/vrw/2020/09090634/1jIxkrgIlEY", "parentPublication": { "id": "proceedings/vrw/2020/6532/0", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vrw/2021/4057/0/405700a607", "title": "Virtual Walking Generator from Omnidirectional Video with Ground-dependent Foot Vibrations", "doi": null, "abstractUrl": "/proceedings-article/vrw/2021/405700a607/1tnWZe0CPwA", "parentPublication": { "id": "proceedings/vrw/2021/4057/0", "title": "2021 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNynJMVB", "title": "2017 IEEE Symposium on Visualization for Cyber Security (VizSec)", "acronym": "vizsec", "groupId": "1810104", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNBKmXmJ", "doi": "10.1109/VIZSEC.2017.8062202", "title": "Adversarial-Playground: A visualization suite showing how adversarial examples fool deep learning", "normalizedTitle": "Adversarial-Playground: A visualization suite showing how adversarial examples fool deep learning", "abstract": "Recent studies have shown that attackers can force deep learning models to misclassify so-called “adversarial examples:” maliciously generated images formed by making imperceptible modifications to pixel values. With growing interest in deep learning for security applications, it is important for security experts and users of machine learning to recognize how learning systems may be attacked. Due to the complex nature of deep learning, it is challenging to understand how deep models can be fooled by adversarial examples. Thus, we present a web-based visualization tool, Adversarial-Playground, to demonstrate the efficacy of common adversarial methods against a convolutional neural network (CNN) system. Adversarial-Playground is educational, modular and interactive. (1) It enables non-experts to compare examples visually and to understand why an adversarial example can fool a CNN-based image classifier. (2) It can help security experts explore more vulnerability of deep learning as a software module. (3) Building an interactive visualization is challenging in this domain due to the large feature space of image classification (generating adversarial examples is slow in general and visualizing images are costly). Through multiple novel design choices, our tool can provide fast and accurate responses to user requests. Empirically, we find that our client-server division strategy reduced the response time by an average of 1.5 seconds per sample. Our other innovation, a faster variant of JSMA evasion algorithm, empirically performed twice as fast as JSMA and yet maintains a comparable evasion rate1.", "abstracts": [ { "abstractType": "Regular", "content": "Recent studies have shown that attackers can force deep learning models to misclassify so-called “adversarial examples:” maliciously generated images formed by making imperceptible modifications to pixel values. With growing interest in deep learning for security applications, it is important for security experts and users of machine learning to recognize how learning systems may be attacked. Due to the complex nature of deep learning, it is challenging to understand how deep models can be fooled by adversarial examples. Thus, we present a web-based visualization tool, Adversarial-Playground, to demonstrate the efficacy of common adversarial methods against a convolutional neural network (CNN) system. Adversarial-Playground is educational, modular and interactive. (1) It enables non-experts to compare examples visually and to understand why an adversarial example can fool a CNN-based image classifier. (2) It can help security experts explore more vulnerability of deep learning as a software module. (3) Building an interactive visualization is challenging in this domain due to the large feature space of image classification (generating adversarial examples is slow in general and visualizing images are costly). Through multiple novel design choices, our tool can provide fast and accurate responses to user requests. Empirically, we find that our client-server division strategy reduced the response time by an average of 1.5 seconds per sample. Our other innovation, a faster variant of JSMA evasion algorithm, empirically performed twice as fast as JSMA and yet maintains a comparable evasion rate1.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Recent studies have shown that attackers can force deep learning models to misclassify so-called “adversarial examples:” maliciously generated images formed by making imperceptible modifications to pixel values. With growing interest in deep learning for security applications, it is important for security experts and users of machine learning to recognize how learning systems may be attacked. Due to the complex nature of deep learning, it is challenging to understand how deep models can be fooled by adversarial examples. Thus, we present a web-based visualization tool, Adversarial-Playground, to demonstrate the efficacy of common adversarial methods against a convolutional neural network (CNN) system. Adversarial-Playground is educational, modular and interactive. (1) It enables non-experts to compare examples visually and to understand why an adversarial example can fool a CNN-based image classifier. (2) It can help security experts explore more vulnerability of deep learning as a software module. (3) Building an interactive visualization is challenging in this domain due to the large feature space of image classification (generating adversarial examples is slow in general and visualizing images are costly). Through multiple novel design choices, our tool can provide fast and accurate responses to user requests. Empirically, we find that our client-server division strategy reduced the response time by an average of 1.5 seconds per sample. Our other innovation, a faster variant of JSMA evasion algorithm, empirically performed twice as fast as JSMA and yet maintains a comparable evasion rate1.", "fno": "08062202", "keywords": [ "Machine Learning", "Servers", "Security", "Tools", "Libraries", "Neural Networks", "Visualization", "I 2 6 Artificial Intelligence Learning Connectionism And Neural Nets", "K 6 5 Management Of Computing And Information Systems Security And Protection Unauthorized Access" ], "authors": [ { "affiliation": "Department of Computer Science, University of Virginia", "fullName": "Andrew P. Norton", "givenName": "Andrew P.", "surname": "Norton", "__typename": "ArticleAuthorType" }, { "affiliation": "Department of Computer Science, University of Virginia", "fullName": "Yanjun Qi", "givenName": "Yanjun", "surname": "Qi", "__typename": "ArticleAuthorType" } ], "idPrefix": "vizsec", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-10-01T00:00:00", "pubType": "proceedings", "pages": "1-4", "year": "2017", "issn": null, "isbn": "978-1-5386-2693-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08062201", "articleId": "12OmNrAv3M3", "__typename": "AdjacentArticleType" }, "next": { "fno": "08062203", "articleId": "12OmNxAlAbr", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccvw/2017/1034/0/1034a751", "title": "Is Deep Learning Safe for Robot Vision? Adversarial Examples Against the iCub Humanoid", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2017/1034a751/12OmNBgQFJf", "parentPublication": { "id": "proceedings/iccvw/2017/1034/0", "title": "2017 IEEE International Conference on Computer Vision Workshop (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2017/1032/0/1032f775", "title": "Adversarial Examples Detection in Deep Networks with Convolutional Filter Statistics", "doi": null, "abstractUrl": "/proceedings-article/iccv/2017/1032f775/12OmNrNh0K2", "parentPublication": { "id": "proceedings/iccv/2017/1032/0", "title": "2017 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/spw/2018/8276/0/634901a036", "title": "Adversarial Examples for Generative Models", "doi": null, "abstractUrl": "/proceedings-article/spw/2018/634901a036/12UTFEaMbF6", "parentPublication": { "id": "proceedings/spw/2018/8276/0", "title": "2018 IEEE Security and Privacy Workshops (SPW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aiam/2021/1732/0/173200a256", "title": "Flow-Pronged Defense against Adversarial Examples", "doi": null, "abstractUrl": "/proceedings-article/aiam/2021/173200a256/1BzTO1f6rVC", "parentPublication": { "id": "proceedings/aiam/2021/1732/0", "title": "2021 3rd International Conference on Artificial Intelligence and Advanced Manufacture (AIAM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dsc/2021/1815/0/181500a272", "title": "Generate Adversarial Examples Combined with Image Entropy Distribution", "doi": null, "abstractUrl": "/proceedings-article/dsc/2021/181500a272/1CuhXNkw0ne", "parentPublication": { "id": "proceedings/dsc/2021/1815/0", "title": "2021 IEEE Sixth International Conference on Data Science in Cyberspace (DSC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/msn/2021/0668/0/066800a303", "title": "Learning Discriminative Features for Adversarial Robustness", "doi": null, "abstractUrl": "/proceedings-article/msn/2021/066800a303/1CxzxqMY2ys", "parentPublication": { "id": "proceedings/msn/2021/0668/0", "title": "2021 17th International Conference on Mobility, Sensing and Networking (MSN)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900a178", "title": "Exploring Robustness Connection between Artificial and Natural Adversarial Examples", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900a178/1G575ZO4Tx6", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300e772", "title": "Semantic Adversarial Attacks: Parametric Transformations That Fool Deep Classifiers", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300e772/1hQqirRVLeU", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csci/2019/5584/0/558400a371", "title": "Adversarial Examples in Arabic", "doi": null, "abstractUrl": "/proceedings-article/csci/2019/558400a371/1jdE0d0dfZ6", "parentPublication": { "id": "proceedings/csci/2019/5584/0", "title": "2019 International Conference on Computational Science and Computational Intelligence (CSCI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800e665", "title": "Modeling Biological Immunity to Adversarial Examples", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800e665/1m3ofMfQove", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxG1yTI", "title": "2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA)", "acronym": "icmla", "groupId": "1001544", "volume": "0", "displayVolume": "0", "year": "2016", "__typename": "ProceedingType" }, "article": { "id": "12OmNqHItuw", "doi": "10.1109/ICMLA.2016.0020", "title": "Assessing Threat of Adversarial Examples on Deep Neural Networks", "normalizedTitle": "Assessing Threat of Adversarial Examples on Deep Neural Networks", "abstract": "Deep neural networks are facing a potential security threat from adversarial examples, inputs that look normal but cause an incorrect classification by the deep neural network. For example, the proposed threat could result in hand-written digits on a scanned check being incorrectly classified but looking normal when humans see them. This research assesses the extent to which adversarial examples pose a security threat, when one considers the normal image acquisition process. This process is mimicked by simulating the transformations that normally occur in of acquiring the image in a real world application, such as using a scanner to acquire digits for a check amount or using a camera in an autonomous car. These small transformations negate the effect of the carefully crafted perturbations of adversarial examples, resulting in a correct classification by the deep neural network. Thus just acquiring the image decreases the potential impact of the proposed security threat. We also show that the already widely used process of averaging over multiple crops neutralizes most adversarial examples. Normal preprocessing, such as text binarization, almost completely neutralizes adversarial examples. This is the first paper to show that for text driven classification, adversarial examples are an academic curiosity, not a security threat.", "abstracts": [ { "abstractType": "Regular", "content": "Deep neural networks are facing a potential security threat from adversarial examples, inputs that look normal but cause an incorrect classification by the deep neural network. For example, the proposed threat could result in hand-written digits on a scanned check being incorrectly classified but looking normal when humans see them. This research assesses the extent to which adversarial examples pose a security threat, when one considers the normal image acquisition process. This process is mimicked by simulating the transformations that normally occur in of acquiring the image in a real world application, such as using a scanner to acquire digits for a check amount or using a camera in an autonomous car. These small transformations negate the effect of the carefully crafted perturbations of adversarial examples, resulting in a correct classification by the deep neural network. Thus just acquiring the image decreases the potential impact of the proposed security threat. We also show that the already widely used process of averaging over multiple crops neutralizes most adversarial examples. Normal preprocessing, such as text binarization, almost completely neutralizes adversarial examples. This is the first paper to show that for text driven classification, adversarial examples are an academic curiosity, not a security threat.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Deep neural networks are facing a potential security threat from adversarial examples, inputs that look normal but cause an incorrect classification by the deep neural network. For example, the proposed threat could result in hand-written digits on a scanned check being incorrectly classified but looking normal when humans see them. This research assesses the extent to which adversarial examples pose a security threat, when one considers the normal image acquisition process. This process is mimicked by simulating the transformations that normally occur in of acquiring the image in a real world application, such as using a scanner to acquire digits for a check amount or using a camera in an autonomous car. These small transformations negate the effect of the carefully crafted perturbations of adversarial examples, resulting in a correct classification by the deep neural network. Thus just acquiring the image decreases the potential impact of the proposed security threat. We also show that the already widely used process of averaging over multiple crops neutralizes most adversarial examples. Normal preprocessing, such as text binarization, almost completely neutralizes adversarial examples. This is the first paper to show that for text driven classification, adversarial examples are an academic curiosity, not a security threat.", "fno": "07838124", "keywords": [ "Handwritten Character Recognition", "Image Classification", "Neural Nets", "Optical Character Recognition", "Security Of Data", "Text Analysis", "Deep Neural Networks", "Threat Assessment", "Security Threat", "Hand Written Digits", "Normal Image Acquisition", "Image Classification", "Text Driven Classification", "Adversarial Examples", "Neural Networks", "Security", "Agriculture", "Training", "Machine Learning", "MIMI Cs", "Cameras" ], "authors": [ { "affiliation": null, "fullName": "Abigail Graese", "givenName": "Abigail", "surname": "Graese", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Andras Rozsa", "givenName": "Andras", "surname": "Rozsa", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Terrance E. Boult", "givenName": "Terrance E.", "surname": "Boult", "__typename": "ArticleAuthorType" } ], "idPrefix": "icmla", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2016-12-01T00:00:00", "pubType": "proceedings", "pages": "69-74", "year": "2016", "issn": null, "isbn": "978-1-5090-6167-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07838123", "articleId": "12OmNx76TGW", "__typename": "AdjacentArticleType" }, "next": { "fno": "07838125", "articleId": "12OmNwDACuJ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vizsec/2017/2693/0/08062202", "title": "Adversarial-Playground: A visualization suite showing how adversarial examples fool deep learning", "doi": null, "abstractUrl": "/proceedings-article/vizsec/2017/08062202/12OmNBKmXmJ", "parentPublication": { "id": "proceedings/vizsec/2017/2693/0", "title": "2017 IEEE Symposium on Visualization for Cyber Security (VizSec)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2017/1034/0/1034a751", "title": "Is Deep Learning Safe for Robot Vision? Adversarial Examples Against the iCub Humanoid", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2017/1034a751/12OmNBgQFJf", "parentPublication": { "id": "proceedings/iccvw/2017/1034/0", "title": "2017 IEEE International Conference on Computer Vision Workshop (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2017/1032/0/1032f775", "title": "Adversarial Examples Detection in Deep Networks with Convolutional Filter Statistics", "doi": null, "abstractUrl": "/proceedings-article/iccv/2017/1032f775/12OmNrNh0K2", "parentPublication": { "id": "proceedings/iccv/2017/1032/0", "title": "2017 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aiam/2021/1732/0/173200a256", "title": "Flow-Pronged Defense against Adversarial Examples", "doi": null, "abstractUrl": "/proceedings-article/aiam/2021/173200a256/1BzTO1f6rVC", "parentPublication": { "id": "proceedings/aiam/2021/1732/0", "title": "2021 3rd International Conference on Artificial Intelligence and Advanced Manufacture (AIAM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bigcom/2022/7384/0/738400a038", "title": "Adversarial Examples Detection of Electromagnetic Signal Based on GAN", "doi": null, "abstractUrl": "/proceedings-article/bigcom/2022/738400a038/1LFKHovEeT6", "parentPublication": { "id": "proceedings/bigcom/2022/7384/0", "title": "2022 8th International Conference on Big Data Computing and Communications (BigCom)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2020/9360/0/09150824", "title": "Robust Assessment of Real-World Adversarial Examples", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2020/09150824/1lPHdlpDH3y", "parentPublication": { "id": "proceedings/cvprw/2020/9360/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dsn-w/2020/7263/0/09151460", "title": "On The Generation of Unrestricted Adversarial Examples", "doi": null, "abstractUrl": "/proceedings-article/dsn-w/2020/09151460/1lRm1DTaWiY", "parentPublication": { "id": "proceedings/dsn-w/2020/7263/0", "title": "2020 50th Annual IEEE/IFIP International Conference on Dependable Systems and Networks Workshops (DSN-W)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800a816", "title": "Adversarial Examples Improve Image Recognition", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800a816/1m3nSN6JtMk", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2021/3864/0/09428316", "title": "Undetectable Adversarial Examples Based on Microscopical Regularization", "doi": null, "abstractUrl": "/proceedings-article/icme/2021/09428316/1uilHXBjdKg", "parentPublication": { "id": "proceedings/icme/2021/3864/0", "title": "2021 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tq/2022/06/09591317", "title": "Generating Adversarial Examples With Distance Constrained Adversarial Imitation Networks", "doi": null, "abstractUrl": "/journal/tq/2022/06/09591317/1y2FEYRs5qM", "parentPublication": { "id": "trans/tq", "title": "IEEE Transactions on Dependable and Secure Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzayNEG", "title": "2017 IEEE Symposium on Security and Privacy (SP)", "acronym": "sp", "groupId": "1000646", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNviHK8t", "doi": "10.1109/SP.2017.49", "title": "Towards Evaluating the Robustness of Neural Networks", "normalizedTitle": "Towards Evaluating the Robustness of Neural Networks", "abstract": "Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples: given an input x and any target classification t, it is possible to find a new input x' that is similar to x but classified as t. This makes it difficult to apply neural networks in security-critical areas. Defensive distillation is a recently proposed approach that can take an arbitrary neural network, and increase its robustness, reducing the success rate of current attacks' ability to find adversarial examples from 95% to 0.5%. In this paper, we demonstrate that defensive distillation does not significantly increase the robustness of neural networks by introducing three new attack algorithms that are successful on both distilled and undistilled neural networks with 100% probability. Our attacks are tailored to three distance metrics used previously in the literature, and when compared to previous adversarial example generation algorithms, our attacks are often much more effective (and never worse). Furthermore, we propose using high-confidence adversarial examples in a simple transferability test we show can also be used to break defensive distillation. We hope our attacks will be used as a benchmark in future defense attempts to create neural networks that resist adversarial examples.", "abstracts": [ { "abstractType": "Regular", "content": "Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples: given an input x and any target classification t, it is possible to find a new input x' that is similar to x but classified as t. This makes it difficult to apply neural networks in security-critical areas. Defensive distillation is a recently proposed approach that can take an arbitrary neural network, and increase its robustness, reducing the success rate of current attacks' ability to find adversarial examples from 95% to 0.5%. In this paper, we demonstrate that defensive distillation does not significantly increase the robustness of neural networks by introducing three new attack algorithms that are successful on both distilled and undistilled neural networks with 100% probability. Our attacks are tailored to three distance metrics used previously in the literature, and when compared to previous adversarial example generation algorithms, our attacks are often much more effective (and never worse). Furthermore, we propose using high-confidence adversarial examples in a simple transferability test we show can also be used to break defensive distillation. We hope our attacks will be used as a benchmark in future defense attempts to create neural networks that resist adversarial examples.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples: given an input x and any target classification t, it is possible to find a new input x' that is similar to x but classified as t. This makes it difficult to apply neural networks in security-critical areas. Defensive distillation is a recently proposed approach that can take an arbitrary neural network, and increase its robustness, reducing the success rate of current attacks' ability to find adversarial examples from 95% to 0.5%. In this paper, we demonstrate that defensive distillation does not significantly increase the robustness of neural networks by introducing three new attack algorithms that are successful on both distilled and undistilled neural networks with 100% probability. Our attacks are tailored to three distance metrics used previously in the literature, and when compared to previous adversarial example generation algorithms, our attacks are often much more effective (and never worse). Furthermore, we propose using high-confidence adversarial examples in a simple transferability test we show can also be used to break defensive distillation. We hope our attacks will be used as a benchmark in future defense attempts to create neural networks that resist adversarial examples.", "fno": "07958570", "keywords": [ "Neural Nets", "Security Of Data", "Neural Networks", "Machine Learning", "Defensive Distillation", "Attack Algorithms", "Distance Metrics", "High Confidence Adversarial Examples", "Transferability Test", "Neural Networks", "Robustness", "Measurement", "Speech Recognition", "Security", "Malware", "Resists" ], "authors": [ { "affiliation": null, "fullName": "Nicholas Carlini", "givenName": "Nicholas", "surname": "Carlini", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "David Wagner", "givenName": "David", "surname": "Wagner", "__typename": "ArticleAuthorType" } ], "idPrefix": "sp", "isOpenAccess": true, "showRecommendedArticles": true, "showBuyMe": false, "hasPdf": true, "pubDate": "2017-05-01T00:00:00", "pubType": "proceedings", "pages": "39-57", "year": "2017", "issn": "2375-1207", "isbn": "978-1-5090-5533-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07958569", "articleId": "12OmNApLGxh", "__typename": "AdjacentArticleType" }, "next": { "fno": "07958571", "articleId": "12OmNBOlliW", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/sp/2016/0824/0/0824a582", "title": "Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/sp/2016/0824a582/12OmNyv7m5E", "parentPublication": { "id": "proceedings/sp/2016/0824/0", "title": "2016 IEEE Symposium on Security and Privacy (SP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2023/02/09761760", "title": "GradDiv: Adversarial Robustness of Randomized Neural Networks via Gradient Diversity Regularization", "doi": null, "abstractUrl": "/journal/tp/2023/02/09761760/1CKMkj4mEBW", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/csci/2021/5841/0/584100a152", "title": "Evaluating Accuracy and Adversarial Robustness of Quanvolutional Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/csci/2021/584100a152/1EpLMoDANPi", "parentPublication": { "id": "proceedings/csci/2021/5841/0", "title": "2021 International Conference on Computational Science and Computational Intelligence (CSCI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sp/2023/9336/0/933600a094", "title": "SoK: Certified Robustness for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/sp/2023/933600a094/1He7XNZytry", "parentPublication": { "id": "proceedings/sp/2023/9336/0/", "title": "2023 IEEE Symposium on Security and Privacy (SP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2022/6814/0/681400a236", "title": "Continual Learning with Adversarial Training to Enhance Robustness of Image Recognition Models", "doi": null, "abstractUrl": "/proceedings-article/cw/2022/681400a236/1I6RNvmZA4w", "parentPublication": { "id": "proceedings/cw/2022/6814/0", "title": "2022 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/msn/2022/6457/0/645700a821", "title": "Low-power Robustness Learning Framework for Adversarial Attack on Edges", "doi": null, "abstractUrl": "/proceedings-article/msn/2022/645700a821/1LUtH7eEGOY", "parentPublication": { "id": "proceedings/msn/2022/6457/0", "title": "2022 18th International Conference on Mobility, Sensing and Networking (MSN)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tq/2022/02/09159878", "title": "Defending Against Adversarial Attack Towards Deep Neural Networks Via Collaborative Multi-Task Training", "doi": null, "abstractUrl": "/journal/tq/2022/02/09159878/1m3mcOUAHtK", "parentPublication": { "id": "trans/tq", "title": "IEEE Transactions on Dependable and Secure Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/candar/2020/8221/0/822100a221", "title": "Bayes without Bayesian Learning for Resisting Adversarial Attacks", "doi": null, "abstractUrl": "/proceedings-article/candar/2020/822100a221/1sA9anovYqs", "parentPublication": { "id": "proceedings/candar/2020/8221/0", "title": "2020 Eighth International Symposium on Computing and Networking (CANDAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tq/2022/06/09551763", "title": "Towards Certifying the Asymmetric Robustness for Neural Networks: Quantification and Applications", "doi": null, "abstractUrl": "/journal/tq/2022/06/09551763/1xgx5HjLZVC", "parentPublication": { "id": "trans/tq", "title": "IEEE Transactions on Dependable and Secure Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iisa/2021/0032/0/09555552", "title": "Robustness of Compressed Deep Neural Networks with Adversarial Training", "doi": null, "abstractUrl": "/proceedings-article/iisa/2021/09555552/1xxct4wMpi0", "parentPublication": { "id": "proceedings/iisa/2021/0032/0", "title": "2021 12th International Conference on Information, Intelligence, Systems & Applications (IISA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNykCcdo", "title": "2018 IEEE Pacific Visualization Symposium (PacificVis)", "acronym": "pacificvis", "groupId": "1001657", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "12OmNwDACu7", "doi": "10.1109/PacificVis.2018.00031", "title": "Visualizing Deep Neural Networks for Text Analytics", "normalizedTitle": "Visualizing Deep Neural Networks for Text Analytics", "abstract": "Deep neural networks (DNNs) have made tremendous progress in many different areas in recent years. How these networks function internally, however, is often not well understood. Advances in under-standing DNNs will benefit and accelerate the development of the field. We present TNNVis, a visualization system that supports un-derstanding of deep neural networks specifically designed to analyze text. TNNVis focuses on DNNs composed of fully connected and convolutional layers. It integrates visual encodings and interaction techniques chosen specifically for our tasks. The tool allows users to: (1) visually explore DNN models with arbitrary input using a combination of node-link diagrams and matrix representation; (2) quickly identify activation values, weights, and feature map patterns within a network; (3) flexibly focus on visual information of interest with threshold, inspection, insight query, and tooltip operations; (4) discover network activation and training patterns through animation; and (5) compare differences between internal activation patterns for different inputs to the DNN. These functions allow neural network researchers to examine their DNN models from new perspectives, producing insights on how these models function. Clustering and summarization techniques are employed to support large convolutional and fully connected layers. Based on several part of speech models with different structure and size, we present multiple use cases where visualization facilitates an understanding of the models.", "abstracts": [ { "abstractType": "Regular", "content": "Deep neural networks (DNNs) have made tremendous progress in many different areas in recent years. How these networks function internally, however, is often not well understood. Advances in under-standing DNNs will benefit and accelerate the development of the field. We present TNNVis, a visualization system that supports un-derstanding of deep neural networks specifically designed to analyze text. TNNVis focuses on DNNs composed of fully connected and convolutional layers. It integrates visual encodings and interaction techniques chosen specifically for our tasks. The tool allows users to: (1) visually explore DNN models with arbitrary input using a combination of node-link diagrams and matrix representation; (2) quickly identify activation values, weights, and feature map patterns within a network; (3) flexibly focus on visual information of interest with threshold, inspection, insight query, and tooltip operations; (4) discover network activation and training patterns through animation; and (5) compare differences between internal activation patterns for different inputs to the DNN. These functions allow neural network researchers to examine their DNN models from new perspectives, producing insights on how these models function. Clustering and summarization techniques are employed to support large convolutional and fully connected layers. Based on several part of speech models with different structure and size, we present multiple use cases where visualization facilitates an understanding of the models.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Deep neural networks (DNNs) have made tremendous progress in many different areas in recent years. How these networks function internally, however, is often not well understood. Advances in under-standing DNNs will benefit and accelerate the development of the field. We present TNNVis, a visualization system that supports un-derstanding of deep neural networks specifically designed to analyze text. TNNVis focuses on DNNs composed of fully connected and convolutional layers. It integrates visual encodings and interaction techniques chosen specifically for our tasks. The tool allows users to: (1) visually explore DNN models with arbitrary input using a combination of node-link diagrams and matrix representation; (2) quickly identify activation values, weights, and feature map patterns within a network; (3) flexibly focus on visual information of interest with threshold, inspection, insight query, and tooltip operations; (4) discover network activation and training patterns through animation; and (5) compare differences between internal activation patterns for different inputs to the DNN. These functions allow neural network researchers to examine their DNN models from new perspectives, producing insights on how these models function. Clustering and summarization techniques are employed to support large convolutional and fully connected layers. Based on several part of speech models with different structure and size, we present multiple use cases where visualization facilitates an understanding of the models.", "fno": "142401a180", "keywords": [ "Data Visualisation", "Feedforward Neural Nets", "Learning Artificial Intelligence", "Natural Language Processing", "Text Analysis", "Interaction Techniques", "DNN Models", "Network Activation", "Training Patterns", "Internal Activation Patterns", "Neural Network Researchers", "Convolutional Connected Layers", "Fully Connected Layers", "Deep Neural Networks", "DN Ns", "Visualization System", "Convolutional Layers", "Visual Encodings", "Text Analytics", "TNN Vis", "Clustering Techniques", "Summarization Techniques", "Speech Models", "Visualization", "Neurons", "Computational Modeling", "Task Analysis", "Convolutional Neural Networks", "Biological Neural Networks", "Information Visualization", "Deep Learning", "Machine Learning", "Visualization Design", "Human Centered Computing" ], "authors": [ { "affiliation": null, "fullName": "Shaoliang Nie", "givenName": "Shaoliang", "surname": "Nie", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Christopher Healey", "givenName": "Christopher", "surname": "Healey", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Kalpesh Padia", "givenName": "Kalpesh", "surname": "Padia", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Samuel Leeman-Munk", "givenName": "Samuel", "surname": "Leeman-Munk", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jordan Benson", "givenName": "Jordan", "surname": "Benson", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Dave Caira", "givenName": "Dave", "surname": "Caira", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Saratendu Sethi", "givenName": "Saratendu", "surname": "Sethi", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ravi Devarajan", "givenName": "Ravi", "surname": "Devarajan", "__typename": "ArticleAuthorType" } ], "idPrefix": "pacificvis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-04-01T00:00:00", "pubType": "proceedings", "pages": "180-189", "year": "2018", "issn": "2165-8773", "isbn": "978-1-5386-1424-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "142401a175", "articleId": "12OmNwtEEMw", "__typename": "AdjacentArticleType" }, "next": { "fno": "142401a190", "articleId": "12OmNqBtj9S", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/big-data/2015/9926/0/07364099", "title": "Genetic deep neural networks using different activation functions for financial data mining", "doi": null, "abstractUrl": "/proceedings-article/big-data/2015/07364099/12OmNAT0mLU", "parentPublication": { "id": "proceedings/big-data/2015/9926/0", "title": "2015 IEEE International Conference on Big Data (Big Data)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/rtss/2018/7908/0/790800a107", "title": "PredJoule: A Timing-Predictable Energy Optimization Framework for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/rtss/2018/790800a107/17D45WLdYQm", "parentPublication": { "id": "proceedings/rtss/2018/7908/0", "title": "2018 IEEE Real-Time Systems Symposium (RTSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ai/5555/01/09802918", "title": "SpinalNet: Deep Neural Network With Gradual Input", "doi": null, "abstractUrl": "/journal/ai/5555/01/09802918/1Eo1ArIO3Ys", "parentPublication": { "id": "trans/ai", "title": "IEEE Transactions on Artificial Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icstw/2019/0888/0/088800a281", "title": "Variable Strength Combinatorial Testing for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icstw/2019/088800a281/1aDTat6htKM", "parentPublication": { "id": "proceedings/icstw/2019/0888/0", "title": "2019 IEEE International Conference on Software Testing, Verification and Validation Workshops (ICSTW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aitest/2019/0492/0/049200a089", "title": "Behavior Pattern-Driven Test Case Selection for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/aitest/2019/049200a089/1aIROMz9QpW", "parentPublication": { "id": "proceedings/aitest/2019/0492/0", "title": "2019 IEEE International Conference On Artificial Intelligence Testing (AITest)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aitest/2019/0492/0/049200a119", "title": "DeepPath: Path-Driven Testing Criteria for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/aitest/2019/049200a119/1aIRP7EiEp2", "parentPublication": { "id": "proceedings/aitest/2019/0492/0", "title": "2019 IEEE International Conference On Artificial Intelligence Testing (AITest)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cse-euc/2019/1664/0/166400a016", "title": "Test4Deep: an Effective White-Box Testing for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/cse-euc/2019/166400a016/1fHkwAfOQKI", "parentPublication": { "id": "proceedings/cse-euc/2019/1664/0", "title": "2019 IEEE International Conference on Computational Science and Engineering (CSE) and IEEE International Conference on Embedded and Ubiquitous Computing (EUC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hpca/2020/6149/0/614900a001", "title": "Deep Learning Acceleration with Neuron-to-Memory Transformation", "doi": null, "abstractUrl": "/proceedings-article/hpca/2020/614900a001/1j9wsaJuOzu", "parentPublication": { "id": "proceedings/hpca/2020/6149/0", "title": "2020 IEEE International Symposium on High Performance Computer Architecture (HPCA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/arith/2020/7120/0/09154496", "title": "A Framework for Semi-Automatic Precision and Accuracy Analysis for Fast and Rigorous Deep Learning", "doi": null, "abstractUrl": "/proceedings-article/arith/2020/09154496/1lZE2UqTVrG", "parentPublication": { "id": "proceedings/arith/2020/7120/0", "title": "2020 IEEE 27th Symposium on Computer Arithmetic (ARITH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/micro/2020/7383/0/738300a738", "title": "DUET: Boosting Deep Neural Network Efficiency on Dual-Module Architecture", "doi": null, "abstractUrl": "/proceedings-article/micro/2020/738300a738/1oFGCNe8bbG", "parentPublication": { "id": "proceedings/micro/2020/7383/0", "title": "2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1BmEezmpGrm", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1BmIAJt1ieI", "doi": "10.1109/ICCV48922.2021.00740", "title": "Towards Robustness of Deep Neural Networks via Regularization", "normalizedTitle": "Towards Robustness of Deep Neural Networks via Regularization", "abstract": "Recent studies have demonstrated the vulnerability of deep neural networks against adversarial examples. In-spired by the observation that adversarial examples often lie outside the natural image data manifold and the intrinsic dimension of image data is much smaller than its pixel space dimension, we propose to embed high-dimensional input images into a low-dimensional space and apply regularization on the embedding space to push the adversarial examples back to the manifold. The proposed framework is called Embedding Regularized Classifier (ER-Classifier), which improves the adversarial robustness of the classifier through embedding regularization. Besides improving classification accuracy against adversarial examples, the framework can be combined with detection methods to detect adversarial examples. Experimental results on several benchmark datasets show that, our proposed framework achieves good performance against strong adversarial at-tack methods.", "abstracts": [ { "abstractType": "Regular", "content": "Recent studies have demonstrated the vulnerability of deep neural networks against adversarial examples. In-spired by the observation that adversarial examples often lie outside the natural image data manifold and the intrinsic dimension of image data is much smaller than its pixel space dimension, we propose to embed high-dimensional input images into a low-dimensional space and apply regularization on the embedding space to push the adversarial examples back to the manifold. The proposed framework is called Embedding Regularized Classifier (ER-Classifier), which improves the adversarial robustness of the classifier through embedding regularization. Besides improving classification accuracy against adversarial examples, the framework can be combined with detection methods to detect adversarial examples. Experimental results on several benchmark datasets show that, our proposed framework achieves good performance against strong adversarial at-tack methods.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Recent studies have demonstrated the vulnerability of deep neural networks against adversarial examples. In-spired by the observation that adversarial examples often lie outside the natural image data manifold and the intrinsic dimension of image data is much smaller than its pixel space dimension, we propose to embed high-dimensional input images into a low-dimensional space and apply regularization on the embedding space to push the adversarial examples back to the manifold. The proposed framework is called Embedding Regularized Classifier (ER-Classifier), which improves the adversarial robustness of the classifier through embedding regularization. Besides improving classification accuracy against adversarial examples, the framework can be combined with detection methods to detect adversarial examples. Experimental results on several benchmark datasets show that, our proposed framework achieves good performance against strong adversarial at-tack methods.", "fno": "281200h476", "keywords": [ "Deep Learning", "Manifolds", "Computer Vision", "Analytical Models", "Computational Modeling", "Neural Networks", "Benchmark Testing", "Adversarial Learning", "Recognition And Classification" ], "authors": [ { "affiliation": "University of North Carolina,Chapel Hill", "fullName": "Yao Li", "givenName": "Yao", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "NEC Labs America,Princeton", "fullName": "Martin Renqiang Min", "givenName": "Martin Renqiang", "surname": "Min", "__typename": "ArticleAuthorType" }, { "affiliation": "University of California,Davis", "fullName": "Thomas Lee", "givenName": "Thomas", "surname": "Lee", "__typename": "ArticleAuthorType" }, { "affiliation": "NEC Labs America,Princeton", "fullName": "Wenchao Yu", "givenName": "Wenchao", "surname": "Yu", "__typename": "ArticleAuthorType" }, { "affiliation": "NEC Labs America,Princeton", "fullName": "Erik Kruus", "givenName": "Erik", "surname": "Kruus", "__typename": "ArticleAuthorType" }, { "affiliation": "University of California,Los Angeles", "fullName": "Wei Wang", "givenName": "Wei", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "University of California,Los Angeles", "fullName": "Cho-Jui Hsieh", "givenName": "Cho-Jui", "surname": "Hsieh", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-10-01T00:00:00", "pubType": "proceedings", "pages": "7476-7485", "year": "2021", "issn": null, "isbn": "978-1-6654-2812-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "281200h466", "articleId": "1BmGhLOFfPO", "__typename": "AdjacentArticleType" }, "next": { "fno": "281200h486", "articleId": "1BmHdwJx5Xq", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/bigcomp/2018/3649/0/364901a066", "title": "ANE: Network Embedding via Adversarial Autoencoders", "doi": null, "abstractUrl": "/proceedings-article/bigcomp/2018/364901a066/12OmNqAU6qu", "parentPublication": { "id": "proceedings/bigcomp/2018/3649/0", "title": "2018 IEEE International Conference on Big Data and Smart Computing (BigComp)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdmw/2018/9288/0/928800a674", "title": "Grading Tumor Malignancy via Deep Bidirectional LSTM on Graph Manifold Encoded Histopathological Image", "doi": null, "abstractUrl": "/proceedings-article/icdmw/2018/928800a674/18jXAUb5as8", "parentPublication": { "id": "proceedings/icdmw/2018/9288/0", "title": "2018 IEEE International Conference on Data Mining Workshops (ICDMW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200h667", "title": "Adversarial Example Detection Using Latent Neighborhood Graph", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200h667/1BmKfwpw7n2", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2023/03/09773978", "title": "Adversarial Robustness Via Fisher-Rao Regularization", "doi": null, "abstractUrl": "/journal/tp/2023/03/09773978/1DjDodGHBhC", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2022/9062/0/09956073", "title": "Curse of co-Dimensionality: Explaining Adversarial Examples by Embedding Geometry of Data Manifold", "doi": null, "abstractUrl": "/proceedings-article/icpr/2022/09956073/1IHqiOsdSPS", "parentPublication": { "id": "proceedings/icpr/2022/9062/0", "title": "2022 26th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2019/9552/0/955200a380", "title": "Query-Free Embedding Attack Against Deep Learning", "doi": null, "abstractUrl": "/proceedings-article/icme/2019/955200a380/1cdOSGtAeAw", "parentPublication": { "id": "proceedings/icme/2019/9552/0", "title": "2019 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300g969", "title": "Disentangling Adversarial Robustness and Generalization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300g969/1gyrB0LeXfO", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300k0668", "title": "Tangent-Normal Adversarial Regularization for Semi-Supervised Learning", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300k0668/1gyscAD4bVS", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/sc/2022/06/09460767", "title": "Defending Adversarial Attacks via Semantic Feature Manipulation", "doi": null, "abstractUrl": "/journal/sc/2022/06/09460767/1uxeOfy1EDS", "parentPublication": { "id": "trans/sc", "title": "IEEE Transactions on Services Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ictai/2021/0898/0/089800b322", "title": "Adversarial Examples Detection and Analysis with Layer-wise Autoencoders", "doi": null, "abstractUrl": "/proceedings-article/ictai/2021/089800b322/1zw5Vv7zSc8", "parentPublication": { "id": "proceedings/ictai/2021/0898/0", "title": "2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1cJ6WsGCn96", "title": "2018 IEEE Conference on Visual Analytics Science and Technology (VAST)", "acronym": "vast", "groupId": "1001630", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "1cJ6WWAb0wo", "doi": "10.1109/VAST.2018.8802509", "title": "Analyzing the Noise Robustness of Deep Neural Networks", "normalizedTitle": "Analyzing the Noise Robustness of Deep Neural Networks", "abstract": "Deep neural networks (DNNs) are vulnerable to maliciously generated adversarial examples. These examples are intentionally designed by making imperceptible perturbations and often mislead a DNN into making an incorrect prediction. This phenomenon means that there is significant risk in applying DNNs to safety-critical applications, such as driverless cars. To address this issue, we present a visual analytics approach to explain the primary cause of the wrong predictions introduced by adversarial examples. The key is to analyze the datapaths of the adversarial examples and compare them with those of the normal examples. A datapath is a group of critical neurons and their connections. To this end, we formulate the datapath extraction as a subset selection problem and approximately solve it based on back-propagation. A multi-level visualization consisting of a segmented DAG (layer level), an Euler diagram (feature map level), and a heat map (neuron level), has been designed to help experts investigate datapaths from the high-level layers to the detailed neuron activations. Two case studies are conducted that demonstrate the promise of our approach in support of explaining the working mechanism of adversarial examples.", "abstracts": [ { "abstractType": "Regular", "content": "Deep neural networks (DNNs) are vulnerable to maliciously generated adversarial examples. These examples are intentionally designed by making imperceptible perturbations and often mislead a DNN into making an incorrect prediction. This phenomenon means that there is significant risk in applying DNNs to safety-critical applications, such as driverless cars. To address this issue, we present a visual analytics approach to explain the primary cause of the wrong predictions introduced by adversarial examples. The key is to analyze the datapaths of the adversarial examples and compare them with those of the normal examples. A datapath is a group of critical neurons and their connections. To this end, we formulate the datapath extraction as a subset selection problem and approximately solve it based on back-propagation. A multi-level visualization consisting of a segmented DAG (layer level), an Euler diagram (feature map level), and a heat map (neuron level), has been designed to help experts investigate datapaths from the high-level layers to the detailed neuron activations. Two case studies are conducted that demonstrate the promise of our approach in support of explaining the working mechanism of adversarial examples.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Deep neural networks (DNNs) are vulnerable to maliciously generated adversarial examples. These examples are intentionally designed by making imperceptible perturbations and often mislead a DNN into making an incorrect prediction. This phenomenon means that there is significant risk in applying DNNs to safety-critical applications, such as driverless cars. To address this issue, we present a visual analytics approach to explain the primary cause of the wrong predictions introduced by adversarial examples. The key is to analyze the datapaths of the adversarial examples and compare them with those of the normal examples. A datapath is a group of critical neurons and their connections. To this end, we formulate the datapath extraction as a subset selection problem and approximately solve it based on back-propagation. A multi-level visualization consisting of a segmented DAG (layer level), an Euler diagram (feature map level), and a heat map (neuron level), has been designed to help experts investigate datapaths from the high-level layers to the detailed neuron activations. Two case studies are conducted that demonstrate the promise of our approach in support of explaining the working mechanism of adversarial examples.", "fno": "08802509", "keywords": [ "Backpropagation", "Data Analysis", "Data Visualisation", "Directed Graphs", "Neural Nets", "DN Ns", "Safety Critical Applications", "Visual Analytics Approach", "Critical Neurons", "Datapath Extraction", "Multilevel Visualization", "Segmented DAG", "Euler Diagram", "Feature Map Level", "Heat Map", "Neuron Level", "High Level Layers", "Noise Robustness", "Deep Neural Networks", "Generated Adversarial Examples", "Subset Selection Problem", "Back Propagation", "Neurons", "Visual Analytics", "Tools", "Feature Extraction", "Biological Neural Networks", "Machine Learning", "Deep Neural Networks", "Robustness", "Adversarial Examples", "Back Propagation", "Multi Level Visualization" ], "authors": [ { "affiliation": "School of Software, Tsinghua University", "fullName": "Mengchen Liu", "givenName": "Mengchen", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "School of Software, Tsinghua University", "fullName": "Shixia Liu", "givenName": "Shixia", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept.of Comp.Sci.Tech., Tsinghua University", "fullName": "Hang Su", "givenName": "Hang", "surname": "Su", "__typename": "ArticleAuthorType" }, { "affiliation": "School of Software, Tsinghua University", "fullName": "Kelei Cao", "givenName": "Kelei", "surname": "Cao", "__typename": "ArticleAuthorType" }, { "affiliation": "Dept.of Comp.Sci.Tech., Tsinghua University", "fullName": "Jun Zhu", "givenName": "Jun", "surname": "Zhu", "__typename": "ArticleAuthorType" } ], "idPrefix": "vast", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-10-01T00:00:00", "pubType": "proceedings", "pages": "60-71", "year": "2018", "issn": null, "isbn": "978-1-5386-6861-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08802454", "articleId": "1cJ6YEzEuQ0", "__typename": "AdjacentArticleType" }, "next": { "fno": "08802415", "articleId": "1cJ6WDNOqXK", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "trans/tg/2018/01/08019879", "title": "Analyzing the Training Processes of Deep Generative Models", "doi": null, "abstractUrl": "/journal/tg/2018/01/08019879/13rRUxAATgA", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200h476", "title": "Towards Robustness of Deep Neural Networks via Regularization", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200h476/1BmIAJt1ieI", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600p5304", "title": "Enhancing Adversarial Robustness for Deep Metric Learning", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600p5304/1H1mUv6EQ80", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aitest/2019/0492/0/049200a089", "title": "Behavior Pattern-Driven Test Case Selection for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/aitest/2019/049200a089/1aIROMz9QpW", "parentPublication": { "id": "proceedings/aitest/2019/0492/0", "title": "2019 IEEE International Conference On Artificial Intelligence Testing (AITest)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08967166", "title": "Analyzing the Noise Robustness of Deep Neural Networks", "doi": null, "abstractUrl": "/journal/tg/2021/07/08967166/1gPjyNWFSgg", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300a588", "title": "Parametric Noise Injection: Trainable Randomness to Improve Deep Neural Network Robustness Against Adversarial Attack", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300a588/1gyrSf60aCA", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2020/6553/0/09093609", "title": "Multi-way Encoding for Robustness", "doi": null, "abstractUrl": "/proceedings-article/wacv/2020/09093609/1jPbs4meEH6", "parentPublication": { "id": "proceedings/wacv/2020/6553/0", "title": "2020 IEEE Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vizsec/2019/3876/0/09161563", "title": "A Visual Analytics Framework for Adversarial Text Generation", "doi": null, "abstractUrl": "/proceedings-article/vizsec/2019/09161563/1m6hHu2C1LG", "parentPublication": { "id": "proceedings/vizsec/2019/3876/0", "title": "2019 IEEE Symposium on Visualization for Cyber Security (VizSec)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icse/2020/7121/0/712100a714", "title": "ReluDiff: Differential Verification of Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icse/2020/712100a714/1pK5gC8SLzW", "parentPublication": { "id": "proceedings/icse/2020/7121/0", "title": "2020 IEEE/ACM 42nd International Conference on Software Engineering (ICSE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ts/2022/11/09599374", "title": "<italic>CAGFuzz:</italic> Coverage-Guided Adversarial Generative Fuzzing Testing for Image-Based Deep Learning Systems", "doi": null, "abstractUrl": "/journal/ts/2022/11/09599374/1yeCbwvfO48", "parentPublication": { "id": "trans/ts", "title": "IEEE Transactions on Software Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1dx8y6v6yZO", "title": "2019 IEEE Security and Privacy Workshops (SPW)", "acronym": "spw", "groupId": "1801671", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1dx8yAChWCc", "doi": "10.1109/SPW.2019.00014", "title": "On the Robustness of Deep K-Nearest Neighbors", "normalizedTitle": "On the Robustness of Deep K-Nearest Neighbors", "abstract": "Despite a large amount of attention on adversarial examples, very few works have demonstrated an effective defense against this threat. We examine Deep k-Nearest Neighbor (DkNN), a proposed defense that combines k-Nearest Neighbor (kNN) and deep learning to improve the model's robustness to adversarial examples. It is challenging to evaluate the robustness of this scheme due to a lack of efficient algorithm for attacking kNN classifiers with large k and high-dimensional data. We propose a heuristic attack that allows us to use gradient descent to find adversarial examples for kNN classifiers, and then apply it to attack the DkNN defense as well. Results suggest that our attack is moderately stronger than any naive attack on kNN and significantly outperforms other attacks on DkNN.", "abstracts": [ { "abstractType": "Regular", "content": "Despite a large amount of attention on adversarial examples, very few works have demonstrated an effective defense against this threat. We examine Deep k-Nearest Neighbor (DkNN), a proposed defense that combines k-Nearest Neighbor (kNN) and deep learning to improve the model's robustness to adversarial examples. It is challenging to evaluate the robustness of this scheme due to a lack of efficient algorithm for attacking kNN classifiers with large k and high-dimensional data. We propose a heuristic attack that allows us to use gradient descent to find adversarial examples for kNN classifiers, and then apply it to attack the DkNN defense as well. Results suggest that our attack is moderately stronger than any naive attack on kNN and significantly outperforms other attacks on DkNN.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Despite a large amount of attention on adversarial examples, very few works have demonstrated an effective defense against this threat. We examine Deep k-Nearest Neighbor (DkNN), a proposed defense that combines k-Nearest Neighbor (kNN) and deep learning to improve the model's robustness to adversarial examples. It is challenging to evaluate the robustness of this scheme due to a lack of efficient algorithm for attacking kNN classifiers with large k and high-dimensional data. We propose a heuristic attack that allows us to use gradient descent to find adversarial examples for kNN classifiers, and then apply it to attack the DkNN defense as well. Results suggest that our attack is moderately stronger than any naive attack on kNN and significantly outperforms other attacks on DkNN.", "fno": "350800a001", "keywords": [ "Gradient Methods", "Learning Artificial Intelligence", "Nearest Neighbour Methods", "Pattern Classification", "Adversarial Examples", "K NN Classifiers", "High Dimensional Data", "Heuristic Attack", "Dk NN Defense", "Naive Attack", "Effective Defense", "Deep Learning", "Deep K Nearest Neighbors", "Robustness", "Training", "Neural Networks", "Perturbation Methods", "Optimization", "Adaptation Models", "Deep Learning", "Adversarial Examples" ], "authors": [ { "affiliation": "University of California, Berkeley", "fullName": "Chawin Sitawarin", "givenName": "Chawin", "surname": "Sitawarin", "__typename": "ArticleAuthorType" }, { "affiliation": "University of California, Berkeley", "fullName": "David Wagner", "givenName": "David", "surname": "Wagner", "__typename": "ArticleAuthorType" } ], "idPrefix": "spw", "isOpenAccess": true, "showRecommendedArticles": true, "showBuyMe": false, "hasPdf": true, "pubDate": "2019-05-01T00:00:00", "pubType": "proceedings", "pages": "1-7", "year": "2019", "issn": null, "isbn": "978-1-7281-3508-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "350800z020", "articleId": "1fTgeH09hss", "__typename": "AdjacentArticleType" }, "next": { "fno": "350800a008", "articleId": "1dx8zJ2NowE", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/big-data/2017/2715/0/08258130", "title": "Noise self-filtering K-nearest neighbors algorithms", "doi": null, "abstractUrl": "/proceedings-article/big-data/2017/08258130/17D45VUZMWM", "parentPublication": { "id": "proceedings/big-data/2017/2715/0", "title": "2017 IEEE International Conference on Big Data (Big Data)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/big-data/2018/5035/0/08621894", "title": "Deep Similarity-Enhanced K Nearest Neighbors", "doi": null, "abstractUrl": "/proceedings-article/big-data/2018/08621894/17D45WGGoMJ", "parentPublication": { "id": "proceedings/big-data/2018/5035/0", "title": "2018 IEEE International Conference on Big Data (Big Data)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2023/04/09858024", "title": "Towards Robust Person Re-Identification by Defending Against Universal Attackers", "doi": null, "abstractUrl": "/journal/tp/2023/04/09858024/1FSY5ieDRXW", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300c834", "title": "K-Nearest Neighbors Hashing", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300c834/1gyrLaDZnO0", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300i759", "title": "Defense Against Adversarial Images Using Web-Scale Nearest-Neighbor Search", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300i759/1gyrN8VpEVa", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2020/9360/0/09150892", "title": "Learning Ordered Top-k Adversarial Attacks via Adversarial Distillation", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2020/09150892/1lPHdSXIeek", "parentPublication": { "id": "proceedings/cvprw/2020/9360/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2020/9360/0/09151033", "title": "DNDNet: Reconfiguring CNN for Adversarial Robustness", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2020/09151033/1lPHfFFgXok", "parentPublication": { "id": "proceedings/cvprw/2020/9360/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800a318", "title": "Benchmarking Adversarial Robustness on Image Classification", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800a318/1m3nTd13uRq", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800o4441", "title": "Detecting Adversarial Samples Using Influence Functions and Nearest Neighbors", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800o4441/1m3okaFW5e8", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/tps-isa/2020/8543/0/854300a236", "title": "Adversarial Deception in Deep Learning: Analysis and Mitigation", "doi": null, "abstractUrl": "/proceedings-article/tps-isa/2020/854300a236/1qyxBGFv2Vi", "parentPublication": { "id": "proceedings/tps-isa/2020/8543/0", "title": "2020 Second IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1jPbbHBGDHq", "title": "2020 IEEE Winter Conference on Applications of Computer Vision (WACV)", "acronym": "wacv", "groupId": "1000040", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1jPbs4meEH6", "doi": "10.1109/WACV45572.2020.9093609", "title": "Multi-way Encoding for Robustness", "normalizedTitle": "Multi-way Encoding for Robustness", "abstract": "Deep models are state-of-the-art for many computer vision tasks including image classification and object detection. However, it has been shown that deep models are vulnerable to adversarial examples. We highlight how one-hot encoding directly contributes to this vulnerability and propose breaking away from this widely-used, but highly-vulnerable mapping. We demonstrate that by leveraging a different output encoding, multi-way encoding, we decorre-late source and target models, making target models more secure. Our approach makes it more difficult for adversaries to find useful gradients for generating adversarial attacks. We present robustness for black-box and white-box attacks on four benchmark datasets: MNIST, CIFAR-10, CIFAR-100, and SVHN. The strength of our approach is also presented in the form of an attack for model watermarking, raising challenges in detecting stolen models.", "abstracts": [ { "abstractType": "Regular", "content": "Deep models are state-of-the-art for many computer vision tasks including image classification and object detection. However, it has been shown that deep models are vulnerable to adversarial examples. We highlight how one-hot encoding directly contributes to this vulnerability and propose breaking away from this widely-used, but highly-vulnerable mapping. We demonstrate that by leveraging a different output encoding, multi-way encoding, we decorre-late source and target models, making target models more secure. Our approach makes it more difficult for adversaries to find useful gradients for generating adversarial attacks. We present robustness for black-box and white-box attacks on four benchmark datasets: MNIST, CIFAR-10, CIFAR-100, and SVHN. The strength of our approach is also presented in the form of an attack for model watermarking, raising challenges in detecting stolen models.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Deep models are state-of-the-art for many computer vision tasks including image classification and object detection. However, it has been shown that deep models are vulnerable to adversarial examples. We highlight how one-hot encoding directly contributes to this vulnerability and propose breaking away from this widely-used, but highly-vulnerable mapping. We demonstrate that by leveraging a different output encoding, multi-way encoding, we decorre-late source and target models, making target models more secure. Our approach makes it more difficult for adversaries to find useful gradients for generating adversarial attacks. We present robustness for black-box and white-box attacks on four benchmark datasets: MNIST, CIFAR-10, CIFAR-100, and SVHN. The strength of our approach is also presented in the form of an attack for model watermarking, raising challenges in detecting stolen models.", "fno": "09093609", "keywords": [ "Computer Vision", "Image Classification", "Image Coding", "Image Watermarking", "Learning Artificial Intelligence", "Neural Nets", "Object Detection", "Object Recognition", "Highly Vulnerable Mapping", "Different Output Encoding", "Multiway Encoding", "Decorre Late Source", "Target Models", "Adversarial Attacks", "White Box Attacks", "CIFAR 100", "Model Watermarking", "Stolen Models", "Deep Models", "One Hot Encoding", "Adversarial Examples", "Object Detection", "Image Classification", "Computer Vision Tasks", "Encoding", "Robustness", "Perturbation Methods", "Training", "Biological System Modeling", "Neurons", "Correlation" ], "authors": [ { "affiliation": "Boston University", "fullName": "Donghyun Kim", "givenName": "Donghyun", "surname": "Kim", "__typename": "ArticleAuthorType" }, { "affiliation": "Boston University", "fullName": "Sarah Adel Bargal", "givenName": "Sarah Adel", "surname": "Bargal", "__typename": "ArticleAuthorType" }, { "affiliation": "Adobe Research", "fullName": "Jianming Zhang", "givenName": "Jianming", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": "Boston University", "fullName": "Stan Sclaroff", "givenName": "Stan", "surname": "Sclaroff", "__typename": "ArticleAuthorType" } ], "idPrefix": "wacv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-03-01T00:00:00", "pubType": "proceedings", "pages": "1341-1349", "year": "2020", "issn": null, "isbn": "978-1-7281-6553-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09093646", "articleId": "1jPbqgaZuwg", "__typename": "AdjacentArticleType" }, "next": { "fno": "09093445", "articleId": "1jPbysdQJzy", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/isvlsi/2018/7099/0/709901a476", "title": "MAT: A Multi-strength Adversarial Training Method to Mitigate Adversarial Attacks", "doi": null, "abstractUrl": "/proceedings-article/isvlsi/2018/709901a476/12OmNxGSm7w", "parentPublication": { "id": "proceedings/isvlsi/2018/7099/0", "title": "2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2018/3788/0/08546023", "title": "Universal Perturbation Generation for Black-box Attack Using Evolutionary Algorithms", "doi": null, "abstractUrl": "/proceedings-article/icpr/2018/08546023/17D45VtKiuf", "parentPublication": { "id": "proceedings/icpr/2018/3788/0", "title": "2018 24th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/crv/2018/6481/0/648101a055", "title": "On the Robustness of Deep Learning Models to Universal Adversarial Attack", "doi": null, "abstractUrl": "/proceedings-article/crv/2018/648101a055/17D45XvMcav", "parentPublication": { "id": "proceedings/crv/2018/6481/0", "title": "2018 15th Conference on Computer and Robot Vision (CRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/big-data/2021/3902/0/09671580", "title": "Evaluating Model Robustness to Adversarial Samples in Network Intrusion Detection", "doi": null, "abstractUrl": "/proceedings-article/big-data/2021/09671580/1A8gSXXPwWY", "parentPublication": { "id": "proceedings/big-data/2021/3902/0", "title": "2021 IEEE International Conference on Big Data (Big Data)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/trustcom/2022/9425/0/942500a184", "title": "Towards Adversarial Robustness with Multidimensional Perturbations via Contrastive Learning", "doi": null, "abstractUrl": "/proceedings-article/trustcom/2022/942500a184/1LFLVGxzNnO", "parentPublication": { "id": "proceedings/trustcom/2022/9425/0", "title": "2022 IEEE International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2019/2506/0/250600a091", "title": "Dropping Pixels for Adversarial Robustness", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2019/250600a091/1iTvtY7nyTK", "parentPublication": { "id": "proceedings/cvprw/2019/2506/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2020/9360/0/09151033", "title": "DNDNet: Reconfiguring CNN for Adversarial Robustness", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2020/09151033/1lPHfFFgXok", "parentPublication": { "id": "proceedings/cvprw/2020/9360/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2020/9360/0/09151035", "title": "MUTE: Inter-class Ambiguity Driven Multi-hot Target Encoding for Deep Neural Network Design", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2020/09151035/1lPHyR5n8Vq", "parentPublication": { "id": "proceedings/cvprw/2020/9360/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800a329", "title": "What It Thinks Is Important Is Important: Robustness Transfers Through Input Gradients", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800a329/1m3nLJxvh4c", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800a318", "title": "Benchmarking Adversarial Robustness on Image Classification", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800a318/1m3nTd13uRq", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1pK5e3DkCcg", "title": "2020 IEEE/ACM 42nd International Conference on Software Engineering (ICSE)", "acronym": "icse", "groupId": "1000691", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1pK5gC8SLzW", "doi": null, "title": "ReluDiff: Differential Verification of Deep Neural Networks", "normalizedTitle": "ReluDiff: Differential Verification of Deep Neural Networks", "abstract": "As deep neural networks are increasingly being deployed in practice, their efficiency has become an important issue. While there are compression techniques for reducing the network's size, energy consumption and computational requirement, they only demonstrate empirically that there is no loss of accuracy, but lack formal guarantees of the compressed network, e.g., in the presence of adversarial examples. Existing verification techniques such as Reluplex, ReluVal, and DeepPoly provide formal guarantees, but they are designed for analyzing a single network instead of the relationship between two networks. To fill the gap, we develop a new method for differential verification of two closely related networks. Our method consists of a fast but approximate forward interval analysis pass followed by a backward pass that iteratively refines the approximation until the desired property is verified. We have two main innovations. During the forward pass, we exploit structural and behavioral similarities of the two networks to more accurately bound the difference between the output neurons of the two networks. Then in the backward pass, we leverage the gradient differences to more accurately compute the most beneficial refinement. Our experiments show that, compared to state-of-the-art verification tools, our method can achieve orders-of-magnitude speedup and prove many more properties than existing tools.", "abstracts": [ { "abstractType": "Regular", "content": "As deep neural networks are increasingly being deployed in practice, their efficiency has become an important issue. While there are compression techniques for reducing the network's size, energy consumption and computational requirement, they only demonstrate empirically that there is no loss of accuracy, but lack formal guarantees of the compressed network, e.g., in the presence of adversarial examples. Existing verification techniques such as Reluplex, ReluVal, and DeepPoly provide formal guarantees, but they are designed for analyzing a single network instead of the relationship between two networks. To fill the gap, we develop a new method for differential verification of two closely related networks. Our method consists of a fast but approximate forward interval analysis pass followed by a backward pass that iteratively refines the approximation until the desired property is verified. We have two main innovations. During the forward pass, we exploit structural and behavioral similarities of the two networks to more accurately bound the difference between the output neurons of the two networks. Then in the backward pass, we leverage the gradient differences to more accurately compute the most beneficial refinement. Our experiments show that, compared to state-of-the-art verification tools, our method can achieve orders-of-magnitude speedup and prove many more properties than existing tools.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "As deep neural networks are increasingly being deployed in practice, their efficiency has become an important issue. While there are compression techniques for reducing the network's size, energy consumption and computational requirement, they only demonstrate empirically that there is no loss of accuracy, but lack formal guarantees of the compressed network, e.g., in the presence of adversarial examples. Existing verification techniques such as Reluplex, ReluVal, and DeepPoly provide formal guarantees, but they are designed for analyzing a single network instead of the relationship between two networks. To fill the gap, we develop a new method for differential verification of two closely related networks. Our method consists of a fast but approximate forward interval analysis pass followed by a backward pass that iteratively refines the approximation until the desired property is verified. We have two main innovations. During the forward pass, we exploit structural and behavioral similarities of the two networks to more accurately bound the difference between the output neurons of the two networks. Then in the backward pass, we leverage the gradient differences to more accurately compute the most beneficial refinement. Our experiments show that, compared to state-of-the-art verification tools, our method can achieve orders-of-magnitude speedup and prove many more properties than existing tools.", "fno": "712100a714", "keywords": [ "Formal Verification", "Iterative Methods", "Neural Nets", "Program Verification", "Differential Verification", "Closely Related Networks", "Fast But Approximate Forward Interval Analysis Pass", "Backward Pass", "Forward Pass", "State Of The Art Verification Tools", "Deep Neural Networks", "Compression Techniques", "Lack Formal Guarantees", "Verification Techniques", "Technological Innovation", "Energy Consumption", "Neurons", "Tools", "Biological Neural Networks", "Formal Verification", "Verification", "Differential Verification", "Deep Neural Networks", "AI Safety" ], "authors": [ { "affiliation": "University of Southern California,Los Angeles,California,USA", "fullName": "Brandon Paulsen", "givenName": "Brandon", "surname": "Paulsen", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Southern California,Los Angeles,California,USA", "fullName": "Jingbo Wang", "givenName": "Jingbo", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Southern California,Los Angeles,California,USA", "fullName": "Chao Wang", "givenName": "Chao", "surname": "Wang", "__typename": "ArticleAuthorType" } ], "idPrefix": "icse", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-10-01T00:00:00", "pubType": "proceedings", "pages": "714-726", "year": "2020", "issn": null, "isbn": "978-1-4503-7121-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "712100a702", "articleId": "1pK5jKfFKpi", "__typename": "AdjacentArticleType" }, "next": { "fno": "712100a727", "articleId": "1pK5lUZaXjW", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iscc/2016/0679/0/07543791", "title": "SDN verification plane for consistency establishment", "doi": null, "abstractUrl": "/proceedings-article/iscc/2016/07543791/12OmNButq3h", "parentPublication": { "id": "proceedings/iscc/2016/0679/0", "title": "2016 IEEE Symposium on Computers and Communication (ISCC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sec/2008/3348/0/3348a388", "title": "Review on Spacecraft Formal System Design Verification", "doi": null, "abstractUrl": "/proceedings-article/sec/2008/3348a388/12OmNCdk2SA", "parentPublication": { "id": "proceedings/sec/2008/3348/0", "title": "Embedded Computing, IEEE International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/tase/2012/4751/0/4751a043", "title": "Formal Verification of Netlog Protocols", "doi": null, "abstractUrl": "/proceedings-article/tase/2012/4751a043/12OmNvlPkAs", "parentPublication": { "id": "proceedings/tase/2012/4751/0", "title": "2012 Sixth International Symposium on Theoretical Aspects of Software Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icsecompanion/2009/3495/0/05071046", "title": "VCC: Contract-based modular verification of concurrent C", "doi": null, "abstractUrl": "/proceedings-article/icsecompanion/2009/05071046/12OmNxwncvV", "parentPublication": { "id": "proceedings/icsecompanion/2009/3495/0", "title": "International Conference on Software Engineering Companion", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/dt/2005/05/d5478", "title": "Verification by the pound", "doi": null, "abstractUrl": "/magazine/dt/2005/05/d5478/13rRUxly99t", "parentPublication": { "id": "mags/dt", "title": "IEEE Design & Test of Computers", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ts/1981/01/01702802", "title": "Consequence Verification of Flowcharts", "doi": null, "abstractUrl": "/journal/ts/1981/01/01702802/13rRUy2YLUn", "parentPublication": { "id": "trans/ts", "title": "IEEE Transactions on Software Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ase/2020/6768/0/676800a784", "title": "NEURODIFF: Scalable Differential Verification of Neural Networks using Fine-Grained Approximation", "doi": null, "abstractUrl": "/proceedings-article/ase/2020/676800a784/1pP3LJ2k5MI", "parentPublication": { "id": "proceedings/ase/2020/6768/0", "title": "2020 35th IEEE/ACM International Conference on Automated Software Engineering (ASE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icse-companion/2021/1219/0/121900a248", "title": "Scalable Quantitative Verification for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icse-companion/2021/121900a248/1sET7oUVUd2", "parentPublication": { "id": "proceedings/icse-companion/2021/1219/0/", "title": "2021 IEEE/ACM 43rd International Conference on Software Engineering: Companion Proceedings (ICSE-Companion)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icse/2021/0296/0/029600a312", "title": "Scalable Quantitative Verification for Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icse/2021/029600a312/1sEXnQDlCso", "parentPublication": { "id": "proceedings/icse/2021/0296/0/", "title": "2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ictai/2021/0898/0/089800a200", "title": "Geometric Path Enumeration for Equivalence Verification of Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/ictai/2021/089800a200/1zw5TwGMPMQ", "parentPublication": { "id": "proceedings/ictai/2021/0898/0", "title": "2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyKJiaV", "title": "Pattern Recognition, International Conference on", "acronym": "icpr", "groupId": "1000545", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNBuL1mn", "doi": "10.1109/ICPR.2010.791", "title": "Keyframe-Guided Automatic Non-linear Video Editing", "normalizedTitle": "Keyframe-Guided Automatic Non-linear Video Editing", "abstract": "We describe a system for generating coherent movies from a collection of unedited videos. The generation process is guided by one or more input keyframes, which determine the content of the generated video. The basic mechanism involves similarity analysis using the histogram intersection function. The function is applied to spatial pyramid histograms computed on the video frames in the collection using Dense SIFT features. A two-directional greedy path finding algorithm is used to select and arrange frames from the collection while maintaining visual similarity, coherence, and continuity. Our system demonstrates promising results on large video collections and is a first step towards increased automation in non-linear video editing.", "abstracts": [ { "abstractType": "Regular", "content": "We describe a system for generating coherent movies from a collection of unedited videos. The generation process is guided by one or more input keyframes, which determine the content of the generated video. The basic mechanism involves similarity analysis using the histogram intersection function. The function is applied to spatial pyramid histograms computed on the video frames in the collection using Dense SIFT features. A two-directional greedy path finding algorithm is used to select and arrange frames from the collection while maintaining visual similarity, coherence, and continuity. Our system demonstrates promising results on large video collections and is a first step towards increased automation in non-linear video editing.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We describe a system for generating coherent movies from a collection of unedited videos. The generation process is guided by one or more input keyframes, which determine the content of the generated video. The basic mechanism involves similarity analysis using the histogram intersection function. The function is applied to spatial pyramid histograms computed on the video frames in the collection using Dense SIFT features. A two-directional greedy path finding algorithm is used to select and arrange frames from the collection while maintaining visual similarity, coherence, and continuity. Our system demonstrates promising results on large video collections and is a first step towards increased automation in non-linear video editing.", "fno": "4109d236", "keywords": [ "Video Processing", "Image Retrieval", "Automatic Video Generation" ], "authors": [ { "affiliation": null, "fullName": "Vaishnavi Rajgopalan", "givenName": "Vaishnavi", "surname": "Rajgopalan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ananth Ranganathan", "givenName": "Ananth", "surname": "Ranganathan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ramgopal Rajagopalan", "givenName": "Ramgopal", "surname": "Rajagopalan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Sudhir P. Mudur", "givenName": "Sudhir P.", "surname": "Mudur", "__typename": "ArticleAuthorType" } ], "idPrefix": "icpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-08-01T00:00:00", "pubType": "proceedings", "pages": "3236-3239", "year": "2010", "issn": "1051-4651", "isbn": "978-0-7695-4109-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4109d232", "articleId": "12OmNrJAdQ2", "__typename": "AdjacentArticleType" }, "next": { "fno": "4109d240", "articleId": "12OmNApLGOR", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/dexa/2009/3763/0/3763a231", "title": "Motion Based Video Classification for SPRITE Generation", "doi": null, "abstractUrl": "/proceedings-article/dexa/2009/3763a231/12OmNAObbJr", "parentPublication": { "id": "proceedings/dexa/2009/3763/0", "title": "2009 20th International Workshop on Database and Expert Systems Application. DEXA 2009", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iih-msp/2008/3278/0/3278a196", "title": "Video Key Frame Extraction Based on Spatial-Temporal Color Distribution", "doi": null, "abstractUrl": "/proceedings-article/iih-msp/2008/3278a196/12OmNAQrYFS", "parentPublication": { "id": "proceedings/iih-msp/2008/3278/0", "title": "2008 Fourth International Conference on Intelligent Information Hiding and Multimedia Signal Processing (IIH-MSP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2008/3381/0/3381a142", "title": "Keyframe Based Video Object Deformation", "doi": null, "abstractUrl": "/proceedings-article/cw/2008/3381a142/12OmNAT0mNU", "parentPublication": { "id": "proceedings/cw/2008/3381/0", "title": "2008 International Conference on Cyberworlds", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cicn/2011/4587/0/4587a073", "title": "A Study on Keyframe Extraction Methods for Video Summary", "doi": null, "abstractUrl": "/proceedings-article/cicn/2011/4587a073/12OmNBDQbnq", "parentPublication": { "id": "proceedings/cicn/2011/4587/0", "title": "Computational Intelligence and Communication Networks, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aina/2004/2051/1/205110171", "title": "Spatial Video Retrieval Based on the Piecewise Method", "doi": null, "abstractUrl": "/proceedings-article/aina/2004/205110171/12OmNrJAdLK", "parentPublication": { "id": "proceedings/aina/2004/2051/1", "title": "18th International Conference on Advanced Information Networking and Applications, 2004. AINA 2004.", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2005/9331/0/01521470", "title": "Discriminative techniques for keyframe selection", "doi": null, "abstractUrl": "/proceedings-article/icme/2005/01521470/12OmNylbotn", "parentPublication": { "id": "proceedings/icme/2005/9331/0", "title": "2005 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icimt/2009/3922/0/3922a211", "title": "Using Color Strings Comparison for Video Frames Retrieval", "doi": null, "abstractUrl": "/proceedings-article/icimt/2009/3922a211/12OmNyrqzBD", "parentPublication": { "id": "proceedings/icimt/2009/3922/0", "title": "Information and Multimedia Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cadgraphics/2011/4497/0/4497a176", "title": "iFrames: A Multi-level Keyframe Extraction and Navigation Tool for Videos", "doi": null, "abstractUrl": "/proceedings-article/cadgraphics/2011/4497a176/12OmNzIUg2q", "parentPublication": { "id": "proceedings/cadgraphics/2011/4497/0", "title": "Computer-Aided Design and Computer Graphics, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/isecs/2009/3643/2/3643b056", "title": "Similarity Retrieval of Video Database Based on 3D Z-string", "doi": null, "abstractUrl": "/proceedings-article/isecs/2009/3643b056/12OmNzwZ6sg", "parentPublication": { "id": null, "title": null, "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdar/2019/3014/0/301400a442", "title": "Text Siamese Network for Video Textual Keyframe Detection", "doi": null, "abstractUrl": "/proceedings-article/icdar/2019/301400a442/1h81xibbDmU", "parentPublication": { "id": "proceedings/icdar/2019/3014/0", "title": "2019 International Conference on Document Analysis and Recognition (ICDAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzlUKpz", "title": "2006 IEEE International Conference on Multimedia and Expo", "acronym": "icme", "groupId": "1000477", "volume": "0", "displayVolume": "0", "year": "2006", "__typename": "ProceedingType" }, "article": { "id": "12OmNqOOrJJ", "doi": "10.1109/ICME.2006.262778", "title": "Video and Audio Editing for Mobile Applications", "normalizedTitle": "Video and Audio Editing for Mobile Applications", "abstract": "Video content creation and consumption have been increasingly available for the masses with the emergence of handheld devices capable of shooting, downloading, and playing videos. Video editing is a natural and necessary operation that is most commonly employed by users for finalizing and organizing their video content. With the constraints in processing power and memory, conventional spatial domain video editing is not a solution for mobile applications. In this paper, we present a complete video editing system for efficiently editing video content on mobile phones using compressed domain editing algorithms. A critical factor from usability point of view is the processing speed of the editing application. We show that with the proposed compressed domain editing system, typical video editing operations can be performed much faster than real-time on today's S60 phones.", "abstracts": [ { "abstractType": "Regular", "content": "Video content creation and consumption have been increasingly available for the masses with the emergence of handheld devices capable of shooting, downloading, and playing videos. Video editing is a natural and necessary operation that is most commonly employed by users for finalizing and organizing their video content. With the constraints in processing power and memory, conventional spatial domain video editing is not a solution for mobile applications. In this paper, we present a complete video editing system for efficiently editing video content on mobile phones using compressed domain editing algorithms. A critical factor from usability point of view is the processing speed of the editing application. We show that with the proposed compressed domain editing system, typical video editing operations can be performed much faster than real-time on today's S60 phones.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Video content creation and consumption have been increasingly available for the masses with the emergence of handheld devices capable of shooting, downloading, and playing videos. Video editing is a natural and necessary operation that is most commonly employed by users for finalizing and organizing their video content. With the constraints in processing power and memory, conventional spatial domain video editing is not a solution for mobile applications. In this paper, we present a complete video editing system for efficiently editing video content on mobile phones using compressed domain editing algorithms. A critical factor from usability point of view is the processing speed of the editing application. We show that with the proposed compressed domain editing system, typical video editing operations can be performed much faster than real-time on today's S60 phones.", "fno": "04036847", "keywords": [], "authors": [ { "affiliation": "Nokia, ari.hourunranta@nokia.com", "fullName": "Ari Hourunranta", "givenName": "Ari", "surname": "Hourunranta", "__typename": "ArticleAuthorType" }, { "affiliation": "Nokia, asad.islam@gmail.com", "fullName": "Asad Islam", "givenName": "Asad", "surname": "Islam", "__typename": "ArticleAuthorType" }, { "affiliation": "Nokia, fehmi.chebil@yahoo.com", "fullName": "Fehmi Chebil", "givenName": "Fehmi", "surname": "Chebil", "__typename": "ArticleAuthorType" } ], "idPrefix": "icme", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2006-07-01T00:00:00", "pubType": "proceedings", "pages": "1305-1308", "year": "2006", "issn": null, "isbn": "1-4244-0366-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "04036846", "articleId": "12OmNzYeAPC", "__typename": "AdjacentArticleType" }, "next": { "fno": "04036848", "articleId": "12OmNCbCrPJ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/msn/2013/5159/0/06726360", "title": "Cloud Based Mobile Video Editing System", "doi": null, "abstractUrl": "/proceedings-article/msn/2013/06726360/12OmNAle6oH", "parentPublication": { "id": "proceedings/msn/2013/5159/0", "title": "2013 Ninth International Conference on Mobile Ad-hoc and Sensor Networks (MSN)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mue/2008/3134/0/3134a282", "title": "Audio-Based Video Editing with Two-Channel Microphone", "doi": null, "abstractUrl": "/proceedings-article/mue/2008/3134a282/12OmNApu5E0", "parentPublication": { "id": "proceedings/mue/2008/3134/0", "title": "Multimedia and Ubiquitous Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mmcs/1997/7819/0/00609603", "title": "Editing techniques for MPEG multiplexed streams", "doi": null, "abstractUrl": "/proceedings-article/mmcs/1997/00609603/12OmNBTs7xA", "parentPublication": { "id": "proceedings/mmcs/1997/7819/0", "title": "Proceedings of IEEE International Conference on Multimedia Computing and Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmcs/1997/7819/0/78190278", "title": "Editing techniques for MPEG multiplexed streams", "doi": null, "abstractUrl": "/proceedings-article/icmcs/1997/78190278/12OmNBv2CmI", "parentPublication": { "id": "proceedings/icmcs/1997/7819/0", "title": "Multimedia Computing and Systems, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/isspit/2007/1834/0/04458055", "title": "Development System for Video and Audio Algorithms", "doi": null, "abstractUrl": "/proceedings-article/isspit/2007/04458055/12OmNCfjetn", "parentPublication": { "id": "proceedings/isspit/2007/1834/0", "title": "2007 IEEE International Symposium on Signal Processing and Information Technology", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icassp/1996/3192/2/00543586", "title": "Nonlinear editing by Generative Video", "doi": null, "abstractUrl": "/proceedings-article/icassp/1996/00543586/12OmNwDSduy", "parentPublication": { "id": "proceedings/icassp/1996/3192/2", "title": "1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2004/8603/1/01394118", "title": "Content based editing of semantic video metadata", "doi": null, "abstractUrl": "/proceedings-article/icme/2004/01394118/12OmNylKB5A", "parentPublication": { "id": "proceedings/icme/2004/8603/1", "title": "2004 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2004/2128/3/212830934", "title": "Gesture Tracking and Recognition for Lecture Video Editing", "doi": null, "abstractUrl": "/proceedings-article/icpr/2004/212830934/12OmNz6iO4y", "parentPublication": { "id": "proceedings/icpr/2004/2128/3", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/mu/2003/02/u2054", "title": "Editing out Video Editing", "doi": null, "abstractUrl": "/magazine/mu/2003/02/u2054/13rRUxlgy0J", "parentPublication": { "id": "mags/mu", "title": "IEEE MultiMedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2021/4899/0/489900b701", "title": "Editing like Humans: A Contextual, Multimodal Framework for Automated Video Editing", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2021/489900b701/1yXsBnK5gYw", "parentPublication": { "id": "proceedings/cvprw/2021/4899/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNC1GueH", "title": "2012 21st International Conference on Pattern Recognition (ICPR 2012)", "acronym": "icpr", "groupId": "1000545", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNvDqsNe", "doi": "", "title": "A structure-based video representation for web video categorization", "normalizedTitle": "A structure-based video representation for web video categorization", "abstract": "In this paper, we propose a video representation that is motivated by the problem of categorizing large web video collections. The representation focuses on capturing the properties of the temporal structure of a video and deploys low-level image features derived from the self-similarity matrix of the video. The bias of the representation towards the temporal structure is based on our hypothesis that this aspect of web video content is among those that are least affected by the enormous web content diversity and that it therefore could provide a reasonable base for comparing videos independent of irrelevant content variations. Although the validation reported in this paper was only a preliminary one, the results already indicate the potential of the proposed video representation to steer towards grouping web videos into meaningful clusters.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, we propose a video representation that is motivated by the problem of categorizing large web video collections. The representation focuses on capturing the properties of the temporal structure of a video and deploys low-level image features derived from the self-similarity matrix of the video. The bias of the representation towards the temporal structure is based on our hypothesis that this aspect of web video content is among those that are least affected by the enormous web content diversity and that it therefore could provide a reasonable base for comparing videos independent of irrelevant content variations. Although the validation reported in this paper was only a preliminary one, the results already indicate the potential of the proposed video representation to steer towards grouping web videos into meaningful clusters.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, we propose a video representation that is motivated by the problem of categorizing large web video collections. The representation focuses on capturing the properties of the temporal structure of a video and deploys low-level image features derived from the self-similarity matrix of the video. The bias of the representation towards the temporal structure is based on our hypothesis that this aspect of web video content is among those that are least affected by the enormous web content diversity and that it therefore could provide a reasonable base for comparing videos independent of irrelevant content variations. Although the validation reported in this paper was only a preliminary one, the results already indicate the potential of the proposed video representation to steer towards grouping web videos into meaningful clusters.", "fno": "06460164", "keywords": [ "Content Based Retrieval", "Feature Extraction", "Image Representation", "Matrix Algebra", "Video Retrieval", "Video Signal Processing", "Web Sites", "Structure Based Video Representation", "Web Video Categorization", "Temporal Video Structure", "Image Features", "Self Similarity Matrix", "Image Color Analysis", "Histograms", "Feature Extraction", "Vectors", "Motion Pictures", "Image Edge Detection", "Interviews" ], "authors": [ { "affiliation": "Delft Multimedia Information Retrieval Lab", "fullName": "Peng Xu", "givenName": "Peng", "surname": "Xu", "__typename": "ArticleAuthorType" }, { "affiliation": "Pattern Recognition Laboratory, Delft University of Technology Mekelweg 4, Delft, The Nethelands", "fullName": "D.M.J. Tax", "givenName": "D.M.J.", "surname": "Tax", "__typename": "ArticleAuthorType" }, { "affiliation": "Delft Multimedia Information Retrieval Lab", "fullName": "Alan Hanjalic", "givenName": "Alan", "surname": "Hanjalic", "__typename": "ArticleAuthorType" } ], "idPrefix": "icpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-11-01T00:00:00", "pubType": "proceedings", "pages": "433-436", "year": "2012", "issn": "1051-4651", "isbn": "978-1-4673-2216-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06460163", "articleId": "12OmNx4gUtf", "__typename": "AdjacentArticleType" }, "next": { "fno": "06460165", "articleId": "12OmNBziBaq", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/ism/2014/4311/0/4311a018", "title": "Cineast: A Multi-feature Sketch-Based Video Retrieval Engine", "doi": null, "abstractUrl": "/proceedings-article/ism/2014/4311a018/12OmNASraWz", "parentPublication": { "id": "proceedings/ism/2014/4311/0", "title": "2014 IEEE International Symposium on Multimedia (ISM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icassp/2004/8484/5/01327209", "title": "Mean shift based video segment representation and applications to replay detection", "doi": null, "abstractUrl": "/proceedings-article/icassp/2004/01327209/12OmNBiygtM", "parentPublication": { "id": "proceedings/icassp/2004/8484/5", "title": "2004 IEEE International Conference on Acoustics, Speech, and Signal Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2012/4711/0/4711a479", "title": "Lagrange-based Video Encoder Optimisation to Enhance Motion Representation in the Compressed-Domain", "doi": null, "abstractUrl": "/proceedings-article/icme/2012/4711a479/12OmNqH9hrf", "parentPublication": { "id": "proceedings/icme/2012/4711/0", "title": "2012 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2016/3568/0/3568a188", "title": "Gameplay Genre Video Classification by Using Mid-Level Video Representation", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2016/3568a188/12OmNqIzhei", "parentPublication": { "id": "proceedings/sibgrapi/2016/3568/0", "title": "2016 29th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2013/5053/0/06474994", "title": "Large-scale web video event classification by use of Fisher Vectors", "doi": null, "abstractUrl": "/proceedings-article/wacv/2013/06474994/12OmNxd4ttg", "parentPublication": { "id": "proceedings/wacv/2013/5053/0", "title": "Applications of Computer Vision, IEEE Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2008/2570/0/04607678", "title": "A novel scheme for video scenes segmentation and semantic representation", "doi": null, "abstractUrl": "/proceedings-article/icme/2008/04607678/12OmNzWx05k", "parentPublication": { "id": "proceedings/icme/2008/2570/0", "title": "2008 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2013/5099/0/5099a226", "title": "A New Method for Static Video Summarization Using Local Descriptors and Video Temporal Segmentation", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2013/5099a226/12OmNzwHvwu", "parentPublication": { "id": "proceedings/sibgrapi/2013/5099/0", "title": "2013 XXVI Conference on Graphics, Patterns and Images", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sitis/2018/9385/0/938500a609", "title": "Leveraging Knowledge Graphs of Movies and Their Content for Web-Scale Analysis", "doi": null, "abstractUrl": "/proceedings-article/sitis/2018/938500a609/19RStzBpuWk", "parentPublication": { "id": "proceedings/sitis/2018/9385/0", "title": "2018 14th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2021/8808/0/09412386", "title": "Automated Whiteboard Lecture Video Summarization by Content Region Detection and Representation", "doi": null, "abstractUrl": "/proceedings-article/icpr/2021/09412386/1tmholo6quA", "parentPublication": { "id": "proceedings/icpr/2021/8808/0", "title": "2020 25th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2021/1838/0/255600a411", "title": "Video Content Representation to Support the Hyper-reality Experience in Virtual Reality", "doi": null, "abstractUrl": "/proceedings-article/vr/2021/255600a411/1tuAwSqh42s", "parentPublication": { "id": "proceedings/vr/2021/1838/0", "title": "2021 IEEE Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBDyAaZ", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNwlHT0C", "doi": "10.1109/ICCV.2015.373", "title": "Multi-cue Structure Preserving MRF for Unconstrained Video Segmentation", "normalizedTitle": "Multi-cue Structure Preserving MRF for Unconstrained Video Segmentation", "abstract": "Video segmentation is a stepping stone to understanding video context. Video segmentation enables one to represent a video by decomposing it into coherent regions which comprise whole or parts of objects. However, the challenge originates from the fact that most of the video segmentation algorithms are based on unsupervised learning due to expensive cost of pixelwise video annotation and intra-class variability within similar unconstrained video classes. We propose a Markov Random Field model for unconstrained video segmentation that relies on tight integration of multiple cues: vertices are defined from contour based superpixels, unary potentials from temporally smooth label likelihood and pairwise potentials from global structure of a video. Multi-cue structure is a breakthrough to extracting coherent object regions for unconstrained videos in absence of supervision. Our experiments on VSB100 dataset show that the proposed model significantly outperforms competing state-of-the-art algorithms. Qualitative analysis illustrates that video segmentation result of the proposed model is consistent with human perception of objects.", "abstracts": [ { "abstractType": "Regular", "content": "Video segmentation is a stepping stone to understanding video context. Video segmentation enables one to represent a video by decomposing it into coherent regions which comprise whole or parts of objects. However, the challenge originates from the fact that most of the video segmentation algorithms are based on unsupervised learning due to expensive cost of pixelwise video annotation and intra-class variability within similar unconstrained video classes. We propose a Markov Random Field model for unconstrained video segmentation that relies on tight integration of multiple cues: vertices are defined from contour based superpixels, unary potentials from temporally smooth label likelihood and pairwise potentials from global structure of a video. Multi-cue structure is a breakthrough to extracting coherent object regions for unconstrained videos in absence of supervision. Our experiments on VSB100 dataset show that the proposed model significantly outperforms competing state-of-the-art algorithms. Qualitative analysis illustrates that video segmentation result of the proposed model is consistent with human perception of objects.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Video segmentation is a stepping stone to understanding video context. Video segmentation enables one to represent a video by decomposing it into coherent regions which comprise whole or parts of objects. However, the challenge originates from the fact that most of the video segmentation algorithms are based on unsupervised learning due to expensive cost of pixelwise video annotation and intra-class variability within similar unconstrained video classes. We propose a Markov Random Field model for unconstrained video segmentation that relies on tight integration of multiple cues: vertices are defined from contour based superpixels, unary potentials from temporally smooth label likelihood and pairwise potentials from global structure of a video. Multi-cue structure is a breakthrough to extracting coherent object regions for unconstrained videos in absence of supervision. Our experiments on VSB100 dataset show that the proposed model significantly outperforms competing state-of-the-art algorithms. Qualitative analysis illustrates that video segmentation result of the proposed model is consistent with human perception of objects.", "fno": "8391d262", "keywords": [ "Motion Segmentation", "Trajectory", "Color", "Proposals", "Image Color Analysis", "Image Edge Detection", "Image Segmentation" ], "authors": [ { "affiliation": null, "fullName": "Saehoon Yi", "givenName": "Saehoon", "surname": "Yi", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Vladimir Pavlovic", "givenName": "Vladimir", "surname": "Pavlovic", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-12-01T00:00:00", "pubType": "proceedings", "pages": "3262-3270", "year": "2015", "issn": "2380-7504", "isbn": "978-1-4673-8391-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "8391d253", "articleId": "12OmNwt5sn2", "__typename": "AdjacentArticleType" }, "next": { "fno": "8391d271", "articleId": "12OmNCu4nd8", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2013/2840/0/2840b777", "title": "Fast Object Segmentation in Unconstrained Video", "doi": null, "abstractUrl": "/proceedings-article/iccv/2013/2840b777/12OmNAYXWEE", "parentPublication": { "id": "proceedings/iccv/2013/2840/0", "title": "2013 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2012/1226/0/06247744", "title": "Exploiting nonlocal spatiotemporal structure for video segmentation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2012/06247744/12OmNCd2rT7", "parentPublication": { "id": "proceedings/cvpr/2012/1226/0", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2016/8851/0/8851a156", "title": "Interactive Segmentation on RGBD Images via Cue Selection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2016/8851a156/12OmNrNh0Jn", "parentPublication": { "id": "proceedings/cvpr/2016/8851/0", "title": "2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdh/2014/4284/0/4284a128", "title": "MRF and CRF Based Image Denoising and Segmentation", "doi": null, "abstractUrl": "/proceedings-article/icdh/2014/4284a128/12OmNwMFMl8", "parentPublication": { "id": "proceedings/icdh/2014/4284/0", "title": "2014 5th International Conference on Digital Home (ICDH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2012/1611/0/06239254", "title": "Video object proposals", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2012/06239254/12OmNxbEtFm", "parentPublication": { "id": "proceedings/cvprw/2012/1611/0", "title": "2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2015/0379/0/0379a266", "title": "A Graph-Based Framework for Video Object Segmentation and Extraction in Feature Space", "doi": null, "abstractUrl": "/proceedings-article/ism/2015/0379a266/12OmNxwENl4", "parentPublication": { "id": "proceedings/ism/2015/0379/0", "title": "2015 IEEE International Symposium on Multimedia (ISM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icicse/2013/5118/0/5118a109", "title": "A Knowledge-Driven Segmentation Method for Ribs in Bone Scintigraphy Using MRF Model", "doi": null, "abstractUrl": "/proceedings-article/icicse/2013/5118a109/12OmNyRPgFc", "parentPublication": { "id": "proceedings/icicse/2013/5118/0", "title": "2013 Seventh International Conference on Internet Computing for Engineering and Science (ICICSE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2018/01/07837719", "title": "Saliency-Aware Video Object Segmentation", "doi": null, "abstractUrl": "/journal/tp/2018/01/07837719/13rRUxCitzQ", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2017/10/07707434", "title": "Video Object Discovery and Co-Segmentation with Extremely Weak Supervision", "doi": null, "abstractUrl": "/journal/tp/2017/10/07707434/13rRUynHukv", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2018/6100/0/610000a630", "title": "Unconstrained Fingerphoto Database", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2018/610000a630/17D45Xh13wn", "parentPublication": { "id": "proceedings/cvprw/2018/6100/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyjLoRf", "title": "Pattern Recognition, International Conference on", "acronym": "icpr", "groupId": "1000545", "volume": "2", "displayVolume": "2", "year": "2002", "__typename": "ProceedingType" }, "article": { "id": "12OmNyOq4SU", "doi": "10.1109/ICPR.2002.1048481", "title": "Video Editing Support System Based on Video Grammar and Content Analysis", "normalizedTitle": "Video Editing Support System Based on Video Grammar and Content Analysis", "abstract": "Video editing is the work to produce the final videos with certain duration by finding and selecting appropriate shots from the material videos and connecting them. In order to produce the excellent videos, this process is generally conducted according to the special rules called \"video grammar\". In this paper, we propose an intelligent support system for the video editing where metadata are extracted automatically and then the video grammars are applied to the extracted metadata.", "abstracts": [ { "abstractType": "Regular", "content": "Video editing is the work to produce the final videos with certain duration by finding and selecting appropriate shots from the material videos and connecting them. In order to produce the excellent videos, this process is generally conducted according to the special rules called \"video grammar\". In this paper, we propose an intelligent support system for the video editing where metadata are extracted automatically and then the video grammars are applied to the extracted metadata.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Video editing is the work to produce the final videos with certain duration by finding and selecting appropriate shots from the material videos and connecting them. In order to produce the excellent videos, this process is generally conducted according to the special rules called \"video grammar\". In this paper, we propose an intelligent support system for the video editing where metadata are extracted automatically and then the video grammars are applied to the extracted metadata.", "fno": "169521031", "keywords": [], "authors": [ { "affiliation": "Ryukoku University", "fullName": "Masahito Kumano", "givenName": "Masahito", "surname": "Kumano", "__typename": "ArticleAuthorType" }, { "affiliation": "Ryukoku University", "fullName": "Yasuo Ariki", "givenName": "Yasuo", "surname": "Ariki", "__typename": "ArticleAuthorType" }, { "affiliation": "Kobe University", "fullName": "Miki Amano", "givenName": "Miki", "surname": "Amano", "__typename": "ArticleAuthorType" }, { "affiliation": "Kobe University", "fullName": "Kuniaki Uehara", "givenName": "Kuniaki", "surname": "Uehara", "__typename": "ArticleAuthorType" }, { "affiliation": "Mainichi Broadcasting System, Inc.", "fullName": "Kenji Shunto", "givenName": "Kenji", "surname": "Shunto", "__typename": "ArticleAuthorType" }, { "affiliation": "Mainichi Broadcasting System, Inc.", "fullName": "Kiyoshi Tsukada", "givenName": "Kiyoshi", "surname": "Tsukada", "__typename": "ArticleAuthorType" } ], "idPrefix": "icpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2002-08-01T00:00:00", "pubType": "proceedings", "pages": "21031", "year": "2002", "issn": "1051-4651", "isbn": "0-7695-1695-X", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "169521025", "articleId": "12OmNAJm0m0", "__typename": "AdjacentArticleType" }, "next": { "fno": "169521037", "articleId": "12OmNAR1aYm", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/mue/2008/3134/0/3134a282", "title": "Audio-Based Video Editing with Two-Channel Microphone", "doi": null, "abstractUrl": "/proceedings-article/mue/2008/3134a282/12OmNApu5E0", "parentPublication": { "id": "proceedings/mue/2008/3134/0", "title": "Multimedia and Ubiquitous Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2006/0366/0/04036847", "title": "Video and Audio Editing for Mobile Applications", "doi": null, "abstractUrl": "/proceedings-article/icme/2006/04036847/12OmNqOOrJJ", "parentPublication": { "id": "proceedings/icme/2006/0366/0", "title": "2006 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2012/4711/0/4711a806", "title": "Automatic Video Editing for Video-Based Interactive Storytelling", "doi": null, "abstractUrl": "/proceedings-article/icme/2012/4711a806/12OmNs0TKI5", "parentPublication": { "id": "proceedings/icme/2012/4711/0", "title": "2012 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cts/2016/2300/0/07870992", "title": "Interoperable Access to Video Content as a Basis for Collaborative Video Editing", "doi": null, "abstractUrl": "/proceedings-article/cts/2016/07870992/12OmNxWLTwK", "parentPublication": { "id": "proceedings/cts/2016/2300/0", "title": "2016 International Conference on Collaboration Technologies and Systems (CTS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2013/2171/0/06746793", "title": "VideoTopic: Content-Based Video Recommendation Using a Topic Model", "doi": null, "abstractUrl": "/proceedings-article/ism/2013/06746793/12OmNyfdOYc", "parentPublication": { "id": "proceedings/ism/2013/2171/0", "title": "2013 IEEE International Symposium on Multimedia (ISM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2004/8603/1/01394118", "title": "Content based editing of semantic video metadata", "doi": null, "abstractUrl": "/proceedings-article/icme/2004/01394118/12OmNylKB5A", "parentPublication": { "id": "proceedings/icme/2004/8603/1", "title": "2004 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2004/2128/3/212830858", "title": "Seamless Video Editing", "doi": null, "abstractUrl": "/proceedings-article/icpr/2004/212830858/12OmNzBwGry", "parentPublication": { "id": "proceedings/icpr/2004/2128/3", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2005/9331/0/01521514", "title": "Video quality analysis for an automated video capturing and editing system for conversation scenes", "doi": null, "abstractUrl": "/proceedings-article/icme/2005/01521514/12OmNzYNN7i", "parentPublication": { "id": "proceedings/icme/2005/9331/0", "title": "2005 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08974422", "title": "Prominent Structures for Video Analysis and Editing", "doi": null, "abstractUrl": "/journal/tg/2021/07/08974422/1gZgXizpprG", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2021/4899/0/489900b701", "title": "Editing like Humans: A Contextual, Multimodal Framework for Automated Video Editing", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2021/489900b701/1yXsBnK5gYw", "parentPublication": { "id": "proceedings/cvprw/2021/4899/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAWH9tO", "title": "Proceedings of International Conference on Acoustics, Speech and Signal Processing (CASSP'02)", "acronym": "icassp", "groupId": "1000002", "volume": "4", "displayVolume": "4", "year": "2002", "__typename": "ProceedingType" }, "article": { "id": "12OmNzGDsGs", "doi": "10.1109/ICASSP.2002.5745432", "title": "A generic video analysis and segmentation system", "normalizedTitle": "A generic video analysis and segmentation system", "abstract": "A generic video analysis system for supervised and unsupervised segmentation is described. The idea behind the presented concept is to integrate different advanced segmentation techniques to obtain a robust, efficient and modular segmentation system for natural video and still images. The system entails several independent modules. Each one of these modules encapsulates a complete video processing technique The intermediate results obtained from each single module are merged and further processed by a set of intelligent rules to achieve a highly accurate final segmentation. The modular structure of the system allows it to be extended continuously and with ease by adding new independent modules. The intermediate segmentation results of newly added modules are linked to the other system results via the rule processor. A user friendly graphical interface (GUI) is also provided. The functionality of the GUI is twofold: it serves as input interface to pass processing parameters to the system and as semi-automatic segmentation tool for user interaction and manually refinement of automatically generated segmentation masks. Selected results obtained with the current version of the video analysis system are reported.", "abstracts": [ { "abstractType": "Regular", "content": "A generic video analysis system for supervised and unsupervised segmentation is described. The idea behind the presented concept is to integrate different advanced segmentation techniques to obtain a robust, efficient and modular segmentation system for natural video and still images. The system entails several independent modules. Each one of these modules encapsulates a complete video processing technique The intermediate results obtained from each single module are merged and further processed by a set of intelligent rules to achieve a highly accurate final segmentation. The modular structure of the system allows it to be extended continuously and with ease by adding new independent modules. The intermediate segmentation results of newly added modules are linked to the other system results via the rule processor. A user friendly graphical interface (GUI) is also provided. The functionality of the GUI is twofold: it serves as input interface to pass processing parameters to the system and as semi-automatic segmentation tool for user interaction and manually refinement of automatically generated segmentation masks. Selected results obtained with the current version of the video analysis system are reported.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "A generic video analysis system for supervised and unsupervised segmentation is described. The idea behind the presented concept is to integrate different advanced segmentation techniques to obtain a robust, efficient and modular segmentation system for natural video and still images. The system entails several independent modules. Each one of these modules encapsulates a complete video processing technique The intermediate results obtained from each single module are merged and further processed by a set of intelligent rules to achieve a highly accurate final segmentation. The modular structure of the system allows it to be extended continuously and with ease by adding new independent modules. The intermediate segmentation results of newly added modules are linked to the other system results via the rule processor. A user friendly graphical interface (GUI) is also provided. The functionality of the GUI is twofold: it serves as input interface to pass processing parameters to the system and as semi-automatic segmentation tool for user interaction and manually refinement of automatically generated segmentation masks. Selected results obtained with the current version of the video analysis system are reported.", "fno": "05745432", "keywords": [ "Image Segmentation", "Image Edge Detection", "Lead", "Image Color Analysis" ], "authors": [ { "affiliation": "Department of Electronic Engineering, Queen Mary, University of London, El 4NS, United Kingdom", "fullName": "Ebroul Izquierdo", "givenName": "Ebroul", "surname": "Izquierdo", "__typename": "ArticleAuthorType" }, { "affiliation": "Department of Electronic Engineering, Queen Mary, University of London, El 4NS, United Kingdom", "fullName": "Jianhui Xia", "givenName": "Jianhui", "surname": "Xia", "__typename": "ArticleAuthorType" }, { "affiliation": "Communication Technology and Information, Processing Institute, University Hannover, Germany", "fullName": "Roland Mech", "givenName": "Roland", "surname": "Mech", "__typename": "ArticleAuthorType" } ], "idPrefix": "icassp", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2002-05-01T00:00:00", "pubType": "proceedings", "pages": "IV-3592-IV-3595", "year": "2002", "issn": "1520-6149", "isbn": null, "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05745431", "articleId": "12OmNzV70G3", "__typename": "AdjacentArticleType" }, "next": { "fno": "05745433", "articleId": "12OmNCctf61", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icpr/2010/4109/0/4109c390", "title": "Color Adjacency Modeling for Improved Image and Video Segmentation", "doi": null, "abstractUrl": "/proceedings-article/icpr/2010/4109c390/12OmNqJ8tvk", "parentPublication": { "id": "proceedings/icpr/2010/4109/0", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2014/4985/0/06836023", "title": "Interactive video segmentation using occlusion boundaries and temporally coherent superpixels", "doi": null, "abstractUrl": "/proceedings-article/wacv/2014/06836023/12OmNwK7o7i", "parentPublication": { "id": "proceedings/wacv/2014/4985/0", "title": "2014 IEEE Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2011/348/0/06011871", "title": "Bilayer video segmentation for videoconferencing applications", "doi": null, "abstractUrl": "/proceedings-article/icme/2011/06011871/12OmNwMXnuo", "parentPublication": { "id": "proceedings/icme/2011/348/0", "title": "2011 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d262", "title": "Multi-cue Structure Preserving MRF for Unconstrained Video Segmentation", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d262/12OmNwlHT0C", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2012/1611/0/06239254", "title": "Video object proposals", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2012/06239254/12OmNxbEtFm", "parentPublication": { "id": "proceedings/cvprw/2012/1611/0", "title": "2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2008/2174/0/04761326", "title": "Monocular video foreground segmentation system", "doi": null, "abstractUrl": "/proceedings-article/icpr/2008/04761326/12OmNyprnw8", "parentPublication": { "id": "proceedings/icpr/2008/2174/0", "title": "ICPR 2008 19th International Conference on Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2008/2570/0/04607750", "title": "Development of a simple free viewpoint video system", "doi": null, "abstractUrl": "/proceedings-article/icme/2008/04607750/12OmNzhELiY", "parentPublication": { "id": "proceedings/icme/2008/2570/0", "title": "2008 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2018/3788/0/08545500", "title": "A New Foreground Segmentation Method for Video Analysis in Different Color Spaces", "doi": null, "abstractUrl": "/proceedings-article/icpr/2018/08545500/17D45VtKitd", "parentPublication": { "id": "proceedings/icpr/2018/3788/0", "title": "2018 24th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2020/6553/0/09093294", "title": "RPM-Net: Robust Pixel-Level Matching Networks for Self-Supervised Video Object Segmentation", "doi": null, "abstractUrl": "/proceedings-article/wacv/2020/09093294/1jPbx94yJ20", "parentPublication": { "id": "proceedings/wacv/2020/6553/0", "title": "2020 IEEE Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800m2943", "title": "Enhancing Generic Segmentation With Learned Region Representations", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800m2943/1m3nD88FtFC", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNB836KO", "title": "2005 IEEE International Conference on Multimedia and Expo", "acronym": "icme", "groupId": "1000477", "volume": "0", "displayVolume": "0", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "12OmNzYNN7i", "doi": "10.1109/ICME.2005.1521514", "title": "Video quality analysis for an automated video capturing and editing system for conversation scenes", "normalizedTitle": "Video quality analysis for an automated video capturing and editing system for conversation scenes", "abstract": "This paper introduces video quality analysis for automated video capture and editing. Previously, we proposed an automated video capture and editing system for conversation scenes. In the capture phase, our system not only produces concurrent video streams with multiple pan-tilt-zoom cameras but also recognizes \"conversation states\" i.e., who is speaking, when someone is nodding, etc. As it is necessary to know the conversation states for the automated editing phase, it is important to clarify how the recognition rate of the conversation attributes affects our editing system with regard to the quality of the resultant videos. In the present study, we analyzed the relationship between the recognition rate of conversation states and the quality of resultant videos through subjective evaluation experiments. The quality scores of the resultant videos were almost the same as the best case in which recognition was done manually, and the recognition rate of our capture system was therefore sufficient.", "abstracts": [ { "abstractType": "Regular", "content": "This paper introduces video quality analysis for automated video capture and editing. Previously, we proposed an automated video capture and editing system for conversation scenes. In the capture phase, our system not only produces concurrent video streams with multiple pan-tilt-zoom cameras but also recognizes \"conversation states\" i.e., who is speaking, when someone is nodding, etc. As it is necessary to know the conversation states for the automated editing phase, it is important to clarify how the recognition rate of the conversation attributes affects our editing system with regard to the quality of the resultant videos. In the present study, we analyzed the relationship between the recognition rate of conversation states and the quality of resultant videos through subjective evaluation experiments. The quality scores of the resultant videos were almost the same as the best case in which recognition was done manually, and the recognition rate of our capture system was therefore sufficient.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper introduces video quality analysis for automated video capture and editing. Previously, we proposed an automated video capture and editing system for conversation scenes. In the capture phase, our system not only produces concurrent video streams with multiple pan-tilt-zoom cameras but also recognizes \"conversation states\" i.e., who is speaking, when someone is nodding, etc. As it is necessary to know the conversation states for the automated editing phase, it is important to clarify how the recognition rate of the conversation attributes affects our editing system with regard to the quality of the resultant videos. In the present study, we analyzed the relationship between the recognition rate of conversation states and the quality of resultant videos through subjective evaluation experiments. The quality scores of the resultant videos were almost the same as the best case in which recognition was done manually, and the recognition rate of our capture system was therefore sufficient.", "fno": "01521514", "keywords": [ "Subjective Evaluation", "Video Quality Analysis", "Automated Video Capturing", "Editing System", "Video Streaming", "Multiple Pan Tilt Zoom Cameras", "Conversation State Recognition" ], "authors": [ { "affiliation": "Graduate Sch. of SIE, Tsukuba Univ., Japan", "fullName": "T. Nishizaki", "givenName": "T.", "surname": "Nishizaki", "__typename": "ArticleAuthorType" }, { "affiliation": "Graduate Sch. of SIE, Tsukuba Univ., Japan", "fullName": "R. Ogata", "givenName": "R.", "surname": "Ogata", "__typename": "ArticleAuthorType" }, { "affiliation": "Graduate Sch. of SIE, Tsukuba Univ., Japan", "fullName": "Y. Kameda", "givenName": "Y.", "surname": "Kameda", "__typename": "ArticleAuthorType" }, { "affiliation": "Graduate Sch. of SIE, Tsukuba Univ., Japan", "fullName": "Y. Ohta", "givenName": "Y.", "surname": "Ohta", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Y. Nakamura", "givenName": "Y.", "surname": "Nakamura", "__typename": "ArticleAuthorType" } ], "idPrefix": "icme", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2005-07-01T00:00:00", "pubType": "proceedings", "pages": "4 pp.", "year": "2005", "issn": null, "isbn": "0-7803-9331-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "01521513", "articleId": "12OmNyuyabK", "__typename": "AdjacentArticleType" }, "next": { "fno": "01521515", "articleId": "12OmNzVoBGX", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/vspets/2005/9424/0/01570913", "title": "Towards intelligent camera networks: a virtual vision approach", "doi": null, "abstractUrl": "/proceedings-article/vspets/2005/01570913/12OmNAWYKFz", "parentPublication": { "id": "proceedings/vspets/2005/9424/0", "title": "Proceedings. 2nd Joint IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance (VS-PETS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/avss/2011/0844/0/06027311", "title": "Multi-tasking smart cameras for intelligent video surveillance systems", "doi": null, "abstractUrl": "/proceedings-article/avss/2011/06027311/12OmNAg7k0q", "parentPublication": { "id": "proceedings/avss/2011/0844/0", "title": "2011 8th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/msn/2013/5159/0/06726360", "title": "Cloud Based Mobile Video Editing System", "doi": null, "abstractUrl": "/proceedings-article/msn/2013/06726360/12OmNAle6oH", "parentPublication": { "id": "proceedings/msn/2013/5159/0", "title": "2013 Ninth International Conference on Mobile Ad-hoc and Sensor Networks (MSN)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mue/2008/3134/0/3134a282", "title": "Audio-Based Video Editing with Two-Channel Microphone", "doi": null, "abstractUrl": "/proceedings-article/mue/2008/3134a282/12OmNApu5E0", "parentPublication": { "id": "proceedings/mue/2008/3134/0", "title": "Multimedia and Ubiquitous Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2010/4109/0/4109d236", "title": "Keyframe-Guided Automatic Non-linear Video Editing", "doi": null, "abstractUrl": "/proceedings-article/icpr/2010/4109d236/12OmNBuL1mn", "parentPublication": { "id": "proceedings/icpr/2010/4109/0", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccis/2010/4270/0/4270a725", "title": "Camera Motion Detection for Conversation Scenes in Movies", "doi": null, "abstractUrl": "/proceedings-article/iccis/2010/4270a725/12OmNrJRPnx", "parentPublication": { "id": "proceedings/iccis/2010/4270/0", "title": "2010 International Conference on Computational and Information Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2012/4711/0/4711a806", "title": "Automatic Video Editing for Video-Based Interactive Storytelling", "doi": null, "abstractUrl": "/proceedings-article/icme/2012/4711a806/12OmNs0TKI5", "parentPublication": { "id": "proceedings/icme/2012/4711/0", "title": "2012 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2004/8603/1/01394118", "title": "Content based editing of semantic video metadata", "doi": null, "abstractUrl": "/proceedings-article/icme/2004/01394118/12OmNylKB5A", "parentPublication": { "id": "proceedings/icme/2004/8603/1", "title": "2004 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/mu/2003/02/u2054", "title": "Editing out Video Editing", "doi": null, "abstractUrl": "/magazine/mu/2003/02/u2054/13rRUxlgy0J", "parentPublication": { "id": "mags/mu", "title": "IEEE MultiMedia", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2021/4899/0/489900b701", "title": "Editing like Humans: A Contextual, Multimodal Framework for Automated Video Editing", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2021/489900b701/1yXsBnK5gYw", "parentPublication": { "id": "proceedings/cvprw/2021/4899/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "17D45VtKiqc", "title": "2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM)", "acronym": "bigmm", "groupId": "1808144", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "17D45Vu1Tzq", "doi": "10.1109/BigMM.2018.8499465", "title": "Automatic Generation of Textual Advertisement for Video Advertising", "normalizedTitle": "Automatic Generation of Textual Advertisement for Video Advertising", "abstract": "With the rapid growth of online videos, video advertising, as a main source of income for video streaming websites, has attracted increasing attention. It is challenging to insert advertisements at suitable positions while preserving comfortable watching experience of users. Existing methods usually insert advertisements at the fixed positions and neglect the variations of scenes, which can extremely reduce the attractiveness of videos due to intrusion to important visual elements. In this paper, we propose a method to automatically generate and embed appealing textual advertisements for online videos. First we estimate the visual significance of the main elements in the video frames via human face localization and saliency detection. Next we design an efficient algorithm to recognize the scene changes with the visual significance map, through which the system can find stable areas in distinct scenes for advertising. At last, a series of aesthetic designing principles are adopted to generate attractive advertisements which are in harmony with the style of video scenes. User studies show that our system can achieve the best user experience compared with the state-of-the-art methods as well as comparable results with commercial advertisements designed by professional designers.", "abstracts": [ { "abstractType": "Regular", "content": "With the rapid growth of online videos, video advertising, as a main source of income for video streaming websites, has attracted increasing attention. It is challenging to insert advertisements at suitable positions while preserving comfortable watching experience of users. Existing methods usually insert advertisements at the fixed positions and neglect the variations of scenes, which can extremely reduce the attractiveness of videos due to intrusion to important visual elements. In this paper, we propose a method to automatically generate and embed appealing textual advertisements for online videos. First we estimate the visual significance of the main elements in the video frames via human face localization and saliency detection. Next we design an efficient algorithm to recognize the scene changes with the visual significance map, through which the system can find stable areas in distinct scenes for advertising. At last, a series of aesthetic designing principles are adopted to generate attractive advertisements which are in harmony with the style of video scenes. User studies show that our system can achieve the best user experience compared with the state-of-the-art methods as well as comparable results with commercial advertisements designed by professional designers.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "With the rapid growth of online videos, video advertising, as a main source of income for video streaming websites, has attracted increasing attention. It is challenging to insert advertisements at suitable positions while preserving comfortable watching experience of users. Existing methods usually insert advertisements at the fixed positions and neglect the variations of scenes, which can extremely reduce the attractiveness of videos due to intrusion to important visual elements. In this paper, we propose a method to automatically generate and embed appealing textual advertisements for online videos. First we estimate the visual significance of the main elements in the video frames via human face localization and saliency detection. Next we design an efficient algorithm to recognize the scene changes with the visual significance map, through which the system can find stable areas in distinct scenes for advertising. At last, a series of aesthetic designing principles are adopted to generate attractive advertisements which are in harmony with the style of video scenes. User studies show that our system can achieve the best user experience compared with the state-of-the-art methods as well as comparable results with commercial advertisements designed by professional designers.", "fno": "08499465", "keywords": [ "Advertising Data Processing", "Face Recognition", "Feature Extraction", "Learning Artificial Intelligence", "Video Signal Processing", "Video Streaming", "Automatic Generation", "Textual Advertisement", "Video Advertising", "Online Videos", "Video Streaming Websites", "Suitable Positions", "Fixed Positions", "Important Visual Elements", "Video Frames", "Visual Significance Map", "Attractive Advertisements", "Video Scenes", "Commercial Advertisements", "Advertising", "Streaming Media", "Visualization", "Optimization", "Saliency Detection", "Color", "Face Detection", "Video Advertising", "Textual Advertisement Generation", "Saliency Detection" ], "authors": [ { "affiliation": "Beijing University of Posts and Telecommunications, Beijing Key Laboratory of Intelligent Telecommunication Software and Multimedia, Beijing, 100876, P. R. China", "fullName": "Yumeng Liang", "givenName": "Yumeng", "surname": "Liang", "__typename": "ArticleAuthorType" }, { "affiliation": "Beijing University of Posts and Telecommunications, Beijing Key Laboratory of Intelligent Telecommunication Software and Multimedia, Beijing, 100876, P. R. China", "fullName": "Wu Liu", "givenName": "Wu", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "Beijing University of Posts and Telecommunications, Beijing Key Laboratory of Intelligent Telecommunication Software and Multimedia, Beijing, 100876, P. R. China", "fullName": "Kun Liu", "givenName": "Kun", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "Beijing University of Posts and Telecommunications, Beijing Key Laboratory of Intelligent Telecommunication Software and Multimedia, Beijing, 100876, P. R. China", "fullName": "Huadong Ma", "givenName": "Huadong", "surname": "Ma", "__typename": "ArticleAuthorType" } ], "idPrefix": "bigmm", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-09-01T00:00:00", "pubType": "proceedings", "pages": "1-5", "year": "2018", "issn": null, "isbn": "978-1-5386-5321-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08499094", "articleId": "17D45Xbl4Qa", "__typename": "AdjacentArticleType" }, "next": { "fno": "08499096", "articleId": "17D45XERml1", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/fg/2015/6026/1/07163153", "title": "To skip or not to skip? A dataset of spontaneous affective response of online advertising (SARA) for audience behavior analysis", "doi": null, "abstractUrl": "/proceedings-article/fg/2015/07163153/12OmNB8kHQX", "parentPublication": { "id": "proceedings/fg/2015/6026/5", "title": "2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icss/2016/2727/0/2727a007", "title": "Clothes Advertising by Targeting Principal Actors in Video", "doi": null, "abstractUrl": "/proceedings-article/icss/2016/2727a007/12OmNC3FGlC", "parentPublication": { "id": "proceedings/icss/2016/2727/0", "title": "2016 9th International Conference on Service Science (ICSS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ism/2015/0379/0/0379a211", "title": "SalAd: A Multimodal Approach for Contextual Video Advertising", "doi": null, "abstractUrl": "/proceedings-article/ism/2015/0379a211/12OmNvo67D4", "parentPublication": { "id": "proceedings/ism/2015/0379/0", "title": "2015 IEEE International Symposium on Multimedia (ISM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/etcs/2010/3987/3/3987c076", "title": "Sensitivity Analysis of Neural Network Parameters for Advertising Images Detection", "doi": null, "abstractUrl": "/proceedings-article/etcs/2010/3987c076/12OmNyUFg0O", "parentPublication": { "id": "proceedings/etcs/2010/3987/3", "title": "Education Technology and Computer Science, International Workshop on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3pgcic/2010/4237/0/4237a505", "title": "Implementation of an Internet Broadcasting System with Video Advertisement Insertion Based on Audience Comments", "doi": null, "abstractUrl": "/proceedings-article/3pgcic/2010/4237a505/12OmNz61dGn", "parentPublication": { "id": "proceedings/3pgcic/2010/4237/0", "title": "P2P, Parallel, Grid, Cloud, and Internet Computing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/nicoint/2018/6909/0/690901a049", "title": "Graph-Based User Interface for Digest Video Creation Focusing on Specific Persons", "doi": null, "abstractUrl": "/proceedings-article/nicoint/2018/690901a049/13bd1eOELLi", "parentPublication": { "id": "proceedings/nicoint/2018/6909/0", "title": "2018 Nicograph International (NicoInt)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmew/2022/7218/0/09859278", "title": "A Low-Cost Virtual 2D Spokes-Character Advertising Framework", "doi": null, "abstractUrl": "/proceedings-article/icmew/2022/09859278/1G4EVRT38wE", "parentPublication": { "id": "proceedings/icmew/2022/7218/0", "title": "2022 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2019/5227/0/522700a186", "title": "Video Audience Analysis using Bayesian Networks and Face Demographics", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2019/522700a186/1fHloYSO0da", "parentPublication": { "id": "proceedings/sibgrapi/2019/5227/0", "title": "2019 32nd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/ta/2022/02/08951232", "title": "Recognition of Advertisement Emotions With Application to Computational Advertising", "doi": null, "abstractUrl": "/journal/ta/2022/02/08951232/1goKZEgDT8I", "parentPublication": { "id": "trans/ta", "title": "IEEE Transactions on Affective Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tk/2023/02/09556599", "title": "Persuade to Click: Context-Aware Persuasion Model for Online Textual Advertisement", "doi": null, "abstractUrl": "/journal/tk/2023/02/09556599/1xlvIuBvXRC", "parentPublication": { "id": "trans/tk", "title": "IEEE Transactions on Knowledge & Data Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzV70Jc", "title": "2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05)", "acronym": "cvpr", "groupId": "1000147", "volume": "2", "displayVolume": "3", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "1htC63xnTYA", "doi": "10.1109/CVPR.2005.369", "title": "Videoshop: a new framework for spatio-temporal video editing in gradient domain", "normalizedTitle": "Videoshop: a new framework for spatio-temporal video editing in gradient domain", "abstract": "Our goal is to develop tools that go beyond frame-constrained manipulation such as resizing, color correction, and simple transitions, and provide object-level operations within frames. Some of our targeted video editing tasks includes transferring a motion picture to a new still picture, importing a moving object into a new background, and compositing two video sequences. The challenges behind this kind of complex video editing tasks lie in two constraints: 1) Spatial consistency: imported objects should blend with the background seamlessly. Hence pixel replacement, which creates noticeable seams, is problematic. 2) Temporal coherency: successive frames should display smooth transitions. Hence frame-by-frame editing, which results in visual flicker, is inappropriate. Our work is aimed at providing an easy-to-use video editing tool that maximally satisfies the spatial and temporal constraints mentioned above and requires minimum user interaction. We propose a new framework for video editing in gradient domain. The spatio-temporal gradient fields of target videos are modified and/or mixed to generate a new gradient field which is usually not integrable. We propose a 3D video integration algorithm, which uses the variational method, to find the potential function whose gradient field is closest to the mixed gradient field in the sense of least squares. The video is reconstructed by solving a 3D Poisson equation. We derive an extension of current 2D gradient technique to 3D space, yielding in a novel video editing framework, which is very different from all current video editing software.", "abstracts": [ { "abstractType": "Regular", "content": "Our goal is to develop tools that go beyond frame-constrained manipulation such as resizing, color correction, and simple transitions, and provide object-level operations within frames. Some of our targeted video editing tasks includes transferring a motion picture to a new still picture, importing a moving object into a new background, and compositing two video sequences. The challenges behind this kind of complex video editing tasks lie in two constraints: 1) Spatial consistency: imported objects should blend with the background seamlessly. Hence pixel replacement, which creates noticeable seams, is problematic. 2) Temporal coherency: successive frames should display smooth transitions. Hence frame-by-frame editing, which results in visual flicker, is inappropriate. Our work is aimed at providing an easy-to-use video editing tool that maximally satisfies the spatial and temporal constraints mentioned above and requires minimum user interaction. We propose a new framework for video editing in gradient domain. The spatio-temporal gradient fields of target videos are modified and/or mixed to generate a new gradient field which is usually not integrable. We propose a 3D video integration algorithm, which uses the variational method, to find the potential function whose gradient field is closest to the mixed gradient field in the sense of least squares. The video is reconstructed by solving a 3D Poisson equation. We derive an extension of current 2D gradient technique to 3D space, yielding in a novel video editing framework, which is very different from all current video editing software.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Our goal is to develop tools that go beyond frame-constrained manipulation such as resizing, color correction, and simple transitions, and provide object-level operations within frames. Some of our targeted video editing tasks includes transferring a motion picture to a new still picture, importing a moving object into a new background, and compositing two video sequences. The challenges behind this kind of complex video editing tasks lie in two constraints: 1) Spatial consistency: imported objects should blend with the background seamlessly. Hence pixel replacement, which creates noticeable seams, is problematic. 2) Temporal coherency: successive frames should display smooth transitions. Hence frame-by-frame editing, which results in visual flicker, is inappropriate. Our work is aimed at providing an easy-to-use video editing tool that maximally satisfies the spatial and temporal constraints mentioned above and requires minimum user interaction. We propose a new framework for video editing in gradient domain. The spatio-temporal gradient fields of target videos are modified and/or mixed to generate a new gradient field which is usually not integrable. We propose a 3D video integration algorithm, which uses the variational method, to find the potential function whose gradient field is closest to the mixed gradient field in the sense of least squares. The video is reconstructed by solving a 3D Poisson equation. We derive an extension of current 2D gradient technique to 3D space, yielding in a novel video editing framework, which is very different from all current video editing software.", "fno": "01467600", "keywords": [ "Video Coding", "Spatiotemporal Phenomena", "Image Sequences", "Gradient Methods", "Variational Techniques", "Image Reconstruction", "Poisson Equation", "Spatio Temporal Video Editing", "Video Sequences", "Spatio Temporal Gradient Fields", "3 D Video Integration Algorithm", "Variational Method", "3 D Poisson Equation", "Videoshop Framework", "Video Compression", "Dynamic Range", "Video Sequences", "Painting", "Cameras", "Apertures", "Laboratories", "Color", "Motion Pictures", "Displays" ], "authors": [ { "affiliation": "Beckman Inst., Illinois Univ., Urbana, IL, USA", "fullName": "H. Wang", "givenName": "H.", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "Beckman Inst., Illinois Univ., Urbana, IL, USA", "fullName": "N. Xu", "givenName": "N.", "surname": "Xu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ramesh Raskar", "givenName": null, "surname": "Ramesh Raskar", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Narendra Ahuja", "givenName": null, "surname": "Narendra Ahuja", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": true, "showRecommendedArticles": true, "showBuyMe": false, "hasPdf": true, "pubDate": "2005-01-01T00:00:00", "pubType": "proceedings", "pages": "1201 vol. 2-1201 vol. 2", "year": "2005", "issn": "1063-6919", "isbn": "0-7695-2372-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "01467599", "articleId": "1htC5K31vfW", "__typename": "AdjacentArticleType" }, "next": { "fno": "01467601", "articleId": "1htC5qBMGY0", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2012/1226/0/109P1C01", "title": "Facial expression editing in video using a temporally-smooth factorization", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2012/109P1C01/12OmNBU1jQU", "parentPublication": { "id": "proceedings/cvpr/2012/1226/0", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2012/4711/0/4711a806", "title": "Automatic Video Editing for Video-Based Interactive Storytelling", "doi": null, "abstractUrl": "/proceedings-article/icme/2012/4711a806/12OmNs0TKI5", "parentPublication": { "id": "proceedings/icme/2012/4711/0", "title": "2012 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icip/1997/8183/1/81831021", "title": "Spatio-temporal video search using the object based video representation", "doi": null, "abstractUrl": "/proceedings-article/icip/1997/81831021/12OmNvjgWH8", "parentPublication": { "id": "proceedings/icip/1997/8183/1", "title": "Image Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dcc/2009/3592/0/3592a382", "title": "Low Complexity Spatio-Temporal Key Frame Encoding for Wyner-Ziv Video Coding", "doi": null, "abstractUrl": "/proceedings-article/dcc/2009/3592a382/12OmNxWuizV", "parentPublication": { "id": "proceedings/dcc/2009/3592/0", "title": "2009 Data Compression Conference. DCC 2009", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccce/2016/2427/0/2427a283", "title": "No-Reference Spatio-temporal Activity Difference PSNR Estimation", "doi": null, "abstractUrl": "/proceedings-article/iccce/2016/2427a283/12OmNyQYtm6", "parentPublication": { "id": "proceedings/iccce/2016/2427/0", "title": "2016 International Conference on Computer and Communication Engineering (ICCCE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2011/0394/0/05995416", "title": "Optimal spatio-temporal path discovery for video event detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2011/05995416/12OmNyXMQ8n", "parentPublication": { "id": "proceedings/cvpr/2011/0394/0", "title": "CVPR 2011", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2004/2128/3/212830858", "title": "Seamless Video Editing", "doi": null, "abstractUrl": "/proceedings-article/icpr/2004/212830858/12OmNzBwGry", "parentPublication": { "id": "proceedings/icpr/2004/2128/3", "title": "Pattern Recognition, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dpvt/2006/2825/0/04155763", "title": "Motion Editing in 3D Video Database", "doi": null, "abstractUrl": "/proceedings-article/3dpvt/2006/04155763/12OmNzwHvtQ", "parentPublication": { "id": "proceedings/3dpvt/2006/2825/0", "title": "Third International Symposium on 3D Data Processing, Visualization, and Transmission (3DPVT'06)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2019/5023/0/502300b513", "title": "Interpretable Spatio-Temporal Attention for Video Action Recognition", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2019/502300b513/1i5mAmpVmHS", "parentPublication": { "id": "proceedings/iccvw/2019/5023/0", "title": "2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900j782", "title": "Spatio-temporal Contrastive Domain Adaptation for Action Recognition", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900j782/1yeKxEVhX7q", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1wzs0vrjyWQ", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "acronym": "cvprw", "groupId": "1001809", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1yXsBnK5gYw", "doi": "10.1109/CVPRW53098.2021.00186", "title": "Editing like Humans: A Contextual, Multimodal Framework for Automated Video Editing", "normalizedTitle": "Editing like Humans: A Contextual, Multimodal Framework for Automated Video Editing", "abstract": "We propose an automated video editing model, which we term contextual and multimodal video editing (CMVE). The model leverages visual and textual metadata describing videos, integrating essential information from both modalities, and uses a learned editing style from a single example video to coherently combine clips. The editing model is useful for tasks such as generating news clip montages and highlight reels given a text query that describes the video storyline. The model exploits the perceptual similarity between video frames, objects in videos and text descriptions to emulate coherent video editing. Amazon Mechanical Turk participants made judgements comparing CMVE to expert human editing. Experimental results showed no significant difference in the CMVE vs human edited video in terms of matching the text query and the level of interest each generates, suggesting CMVE is able to effectively integrate semantic information across visual and textual modalities and create perceptually coherent quality videos typical of human video editors. We publicly release an online demonstration of our method.", "abstracts": [ { "abstractType": "Regular", "content": "We propose an automated video editing model, which we term contextual and multimodal video editing (CMVE). The model leverages visual and textual metadata describing videos, integrating essential information from both modalities, and uses a learned editing style from a single example video to coherently combine clips. The editing model is useful for tasks such as generating news clip montages and highlight reels given a text query that describes the video storyline. The model exploits the perceptual similarity between video frames, objects in videos and text descriptions to emulate coherent video editing. Amazon Mechanical Turk participants made judgements comparing CMVE to expert human editing. Experimental results showed no significant difference in the CMVE vs human edited video in terms of matching the text query and the level of interest each generates, suggesting CMVE is able to effectively integrate semantic information across visual and textual modalities and create perceptually coherent quality videos typical of human video editors. We publicly release an online demonstration of our method.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose an automated video editing model, which we term contextual and multimodal video editing (CMVE). The model leverages visual and textual metadata describing videos, integrating essential information from both modalities, and uses a learned editing style from a single example video to coherently combine clips. The editing model is useful for tasks such as generating news clip montages and highlight reels given a text query that describes the video storyline. The model exploits the perceptual similarity between video frames, objects in videos and text descriptions to emulate coherent video editing. Amazon Mechanical Turk participants made judgements comparing CMVE to expert human editing. Experimental results showed no significant difference in the CMVE vs human edited video in terms of matching the text query and the level of interest each generates, suggesting CMVE is able to effectively integrate semantic information across visual and textual modalities and create perceptually coherent quality videos typical of human video editors. We publicly release an online demonstration of our method.", "fno": "489900b701", "keywords": [ "Training", "Visualization", "Computer Vision", "Conferences", "Semantics", "Metadata", "Pattern Recognition" ], "authors": [ { "affiliation": "Columbia University,Dept. of Biomedical Engineering", "fullName": "Sharath Koorathota", "givenName": "Sharath", "surname": "Koorathota", "__typename": "ArticleAuthorType" }, { "affiliation": "Fovea Inc", "fullName": "Patrick Adelman", "givenName": "Patrick", "surname": "Adelman", "__typename": "ArticleAuthorType" }, { "affiliation": "The Graduate Center, CUNY,Dept. of Psychology", "fullName": "Kelly Cotton", "givenName": "Kelly", "surname": "Cotton", "__typename": "ArticleAuthorType" }, { "affiliation": "Columbia University,Dept. of Biomedical Engineering", "fullName": "Paul Sajda", "givenName": "Paul", "surname": "Sajda", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvprw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-06-01T00:00:00", "pubType": "proceedings", "pages": "1701-1709", "year": "2021", "issn": null, "isbn": "978-1-6654-4899-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1yZ3NuF8ha0", "name": "pcvprw202148990-09522846s1-mm_489900b701.zip", "size": "1.29 MB", "location": "https://www.computer.org/csdl/api/v1/extra/pcvprw202148990-09522846s1-mm_489900b701.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "489900b692", "articleId": "1yZ4y9uUPfi", "__typename": "AdjacentArticleType" }, "next": { "fno": "489900b710", "articleId": "1yVzWoK0xHO", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icme/2006/0366/0/04036847", "title": "Video and Audio Editing for Mobile Applications", "doi": null, "abstractUrl": "/proceedings-article/icme/2006/04036847/12OmNqOOrJJ", "parentPublication": { "id": "proceedings/icme/2006/0366/0", "title": "2006 IEEE International Conference on Multimedia and Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fg/2021/3176/0/09667059", "title": "Emotion Editing in Head Reenactment Videos using Latent Space Manipulation", "doi": null, "abstractUrl": "/proceedings-article/fg/2021/09667059/1A6BJzpRJcs", "parentPublication": { "id": "proceedings/fg/2021/3176/0", "title": "2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500a546", "title": "Multimodal Learning using Optimal Transport for Sarcasm and Humor Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500a546/1B13hIyiVyg", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600k0503", "title": "M<sup>3</sup>L: Language-based Video Editing via Multi-Modal Multi-Level Transformers", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600k0503/1H0MUWHTKHC", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600t9698", "title": "SpaceEdit: Learning a Unified Editing Space for Open-Domain Image Color Editing", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600t9698/1H0NEEhuv7O", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600d605", "title": "Show Me What and Tell Me How: Video Synthesis via Multimodal Conditioning", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600d605/1H1mqE4qrvy", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f573", "title": "Progressive Video Summarization via Multimodal Self-supervised Learning", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f573/1L8qCg1dsic", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08974422", "title": "Prominent Structures for Video Analysis and Editing", "doi": null, "abstractUrl": "/journal/tg/2021/07/08974422/1gZgXizpprG", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2021/4899/0/4.899E165", "title": "Traffic Video Event Retrieval via Text Query using Vehicle Appearance and Motion Attributes", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2021/4.899E165/1yJYuAn90EE", "parentPublication": { "id": "proceedings/cvprw/2021/4899/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2021/4899/0/489900b692", "title": "Private-Shared Disentangled Multimodal VAE for Learning of Latent Representations", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2021/489900b692/1yZ4y9uUPfi", "parentPublication": { "id": "proceedings/cvprw/2021/4899/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxFJXCE", "title": "Energy and Environment Technology, International Conference on", "acronym": "iceet", "groupId": "1003015", "volume": "1", "displayVolume": "1", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNCesr6k", "doi": "10.1109/ICEET.2009.59", "title": "Numerical Simulation of Gas/Solid Flow in a Novel Annular Spouted Bed with Multiple Gas Nozzles", "normalizedTitle": "Numerical Simulation of Gas/Solid Flow in a Novel Annular Spouted Bed with Multiple Gas Nozzles", "abstract": "A novel annular spouted bed with multiple gas nozzles, has been proposed for dryness, pyrolysis, and gasification of coal particulates. It consists of two homocentric upright cylinders with some annularly located spouting gas nozzles between inner and outer cylinders. A three-dimensional Eulerian multiphase model, with closure law according to the kinetic theory of granular flow, was used to simulate the gas/solid flow behaviors in the spouted beds. The simulation results show that numerical simulation is a useful tool to get detailed information about the gas/solid turbulent motions in the novel spouted beds. Along the bed height, the pressure tends to decrease. Particle concentration increases with increasing of distance away from the nozzle. The particle concentration is less for a high spouting gas velocity. The gas velocities at the center axis of nozzle decrease with increasing the static bed height.", "abstracts": [ { "abstractType": "Regular", "content": "A novel annular spouted bed with multiple gas nozzles, has been proposed for dryness, pyrolysis, and gasification of coal particulates. It consists of two homocentric upright cylinders with some annularly located spouting gas nozzles between inner and outer cylinders. A three-dimensional Eulerian multiphase model, with closure law according to the kinetic theory of granular flow, was used to simulate the gas/solid flow behaviors in the spouted beds. The simulation results show that numerical simulation is a useful tool to get detailed information about the gas/solid turbulent motions in the novel spouted beds. Along the bed height, the pressure tends to decrease. Particle concentration increases with increasing of distance away from the nozzle. The particle concentration is less for a high spouting gas velocity. The gas velocities at the center axis of nozzle decrease with increasing the static bed height.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "A novel annular spouted bed with multiple gas nozzles, has been proposed for dryness, pyrolysis, and gasification of coal particulates. It consists of two homocentric upright cylinders with some annularly located spouting gas nozzles between inner and outer cylinders. A three-dimensional Eulerian multiphase model, with closure law according to the kinetic theory of granular flow, was used to simulate the gas/solid flow behaviors in the spouted beds. The simulation results show that numerical simulation is a useful tool to get detailed information about the gas/solid turbulent motions in the novel spouted beds. Along the bed height, the pressure tends to decrease. Particle concentration increases with increasing of distance away from the nozzle. The particle concentration is less for a high spouting gas velocity. The gas velocities at the center axis of nozzle decrease with increasing the static bed height.", "fno": "3819a218", "keywords": [ "Gas Solid Flow", "CFD", "Eulerian Multiphase Model", "Kinetic Theory Of Granular Flow", "Annular Spouted Bed" ], "authors": [ { "affiliation": null, "fullName": "Gong Xi-wu", "givenName": "Gong", "surname": "Xi-wu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Hu Guo-xin", "givenName": "Hu", "surname": "Guo-xin", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zhou Hai-jiang", "givenName": "Zhou", "surname": "Hai-jiang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Shi Qian", "givenName": "Shi", "surname": "Qian", "__typename": "ArticleAuthorType" } ], "idPrefix": "iceet", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-10-01T00:00:00", "pubType": "proceedings", "pages": "218-221", "year": "2009", "issn": null, "isbn": "978-0-7695-3819-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3819a214", "articleId": "12OmNzYwc3C", "__typename": "AdjacentArticleType" }, "next": { "fno": "3819a222", "articleId": "12OmNxXCGHn", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cso/2010/6812/1/05532913", "title": "CFD Simulation on Dense Gas-Solid Flow for Blast Furnace Slag Waste Heat Recovery", "doi": null, "abstractUrl": "/proceedings-article/cso/2010/05532913/12OmNBbaH8H", "parentPublication": { "id": "proceedings/cso/2010/6812/1", "title": "2010 Third International Joint Conference on Computational Science and Optimization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdma/2010/4286/2/4286b423", "title": "Studies of Hydrodynamics of the Fluidized Bed Reactor", "doi": null, "abstractUrl": "/proceedings-article/icdma/2010/4286b423/12OmNrNh0wq", "parentPublication": { "id": "proceedings/icdma/2010/4286/2", "title": "2010 International Conference on Digital Manufacturing & Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceet/2009/3819/3/3819c207", "title": "Applying Numerical Simulation to Analyse the Performance of Nozzles", "doi": null, "abstractUrl": "/proceedings-article/iceet/2009/3819c207/12OmNwErpMB", "parentPublication": { "id": "proceedings/iceet/2009/3819/3", "title": "Energy and Environment Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmtma/2011/4296/2/4296c725", "title": "Numerical Simulation on Spouted Bed Feeding Activated Carbon Particles for Flue Gas Desulfurization", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2011/4296c725/12OmNwsNRf4", "parentPublication": { "id": "proceedings/icmtma/2011/4296/2", "title": "2011 Third International Conference on Measuring Technology and Mechatronics Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icetet/2008/3267/0/3267b094", "title": "CFD Simulations of Heat Transfer in a Bubbling Fluidized Bed for Different Materials", "doi": null, "abstractUrl": "/proceedings-article/icetet/2008/3267b094/12OmNwseERw", "parentPublication": { "id": "proceedings/icetet/2008/3267/0", "title": "Emerging Trends in Engineering & Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdma/2013/5016/0/5016b054", "title": "Research on Gas-Solid Fluidized Bed Agglomeration Early-Warning Technology", "doi": null, "abstractUrl": "/proceedings-article/icdma/2013/5016b054/12OmNx7G5Td", "parentPublication": { "id": "proceedings/icdma/2013/5016/0", "title": "2013 Fourth International Conference on Digital Manufacturing & Automation (ICDMA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceet/2009/3819/3/3819c265", "title": "Desulfurization Characteristics in Double Nozzles Rectangular Spouted Bed with Draft Tube", "doi": null, "abstractUrl": "/proceedings-article/iceet/2009/3819c265/12OmNxwENPC", "parentPublication": { "id": "proceedings/iceet/2009/3819/3", "title": "Energy and Environment Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icece/2010/4031/0/4031d994", "title": "Puffer Characteristic Calculation of a Circuit Breaker Using Fluorocarbon Gas Mixture as an Arc Quenching Medium", "doi": null, "abstractUrl": "/proceedings-article/icece/2010/4031d994/12OmNy3AgtF", "parentPublication": { "id": "proceedings/icece/2010/4031/0", "title": "Electrical and Control Engineering, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceet/2009/3819/1/3819a257", "title": "Pyrolysis Characteristics of Lignite in a Fluidized Bed: Influence of Pyrolysis Temperature", "doi": null, "abstractUrl": "/proceedings-article/iceet/2009/3819a257/12OmNylboNk", "parentPublication": { "id": "proceedings/iceet/2009/3819/1", "title": "Energy and Environment Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icitbs/2021/4854/0/485400a282", "title": "Experimental study on the gas-particle two-phase flow characteristics in a fluidized bed SCR denitrification reactor", "doi": null, "abstractUrl": "/proceedings-article/icitbs/2021/485400a282/1wB71ESwtCU", "parentPublication": { "id": "proceedings/icitbs/2021/4854/0", "title": "2021 International Conference on Intelligent Transportation, Big Data & Smart City (ICITBS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBC8AAG", "title": "2010 Third International Conference on Information and Computing", "acronym": "icic", "groupId": "1002818", "volume": "4", "displayVolume": "4", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNqyDjpQ", "doi": "10.1109/ICIC.2010.271", "title": "Numerical Simulation of the Flow Characteristic with Different Geometrical Jet", "normalizedTitle": "Numerical Simulation of the Flow Characteristic with Different Geometrical Jet", "abstract": "Radial distributions of turbulent kinetic energy k in the tank with different geometrical jet flows was predicted based on CFD method adopted RNG k - ε turbulent model. Some detailed information about turbulence kinetic energy and pressure was obtained. The results showed that at the plane of z=0, pressure did not change significantly and the value was small. With the increase of the distance x, the turbulent kinetic energy enlarged first and then reduced. Turbulent kinetic energy was suddenly reduced above the jet height at x=0. Under the same velocity and the nozzle clearance, the turbulent kinetic energy of flat -bottom was the smallest, and the turbulent kinetic energy of hemispherical-bottom was the largest. Under the same velocity, the pressure of hemispherical-bottom was the largest, wherever the pressure of flat-bottom was the smallest. The results can be theoretical basis for the optimum design and scale-up of the jet mixer.", "abstracts": [ { "abstractType": "Regular", "content": "Radial distributions of turbulent kinetic energy k in the tank with different geometrical jet flows was predicted based on CFD method adopted RNG k - ε turbulent model. Some detailed information about turbulence kinetic energy and pressure was obtained. The results showed that at the plane of z=0, pressure did not change significantly and the value was small. With the increase of the distance x, the turbulent kinetic energy enlarged first and then reduced. Turbulent kinetic energy was suddenly reduced above the jet height at x=0. Under the same velocity and the nozzle clearance, the turbulent kinetic energy of flat -bottom was the smallest, and the turbulent kinetic energy of hemispherical-bottom was the largest. Under the same velocity, the pressure of hemispherical-bottom was the largest, wherever the pressure of flat-bottom was the smallest. The results can be theoretical basis for the optimum design and scale-up of the jet mixer.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Radial distributions of turbulent kinetic energy k in the tank with different geometrical jet flows was predicted based on CFD method adopted RNG k - ε turbulent model. Some detailed information about turbulence kinetic energy and pressure was obtained. The results showed that at the plane of z=0, pressure did not change significantly and the value was small. With the increase of the distance x, the turbulent kinetic energy enlarged first and then reduced. Turbulent kinetic energy was suddenly reduced above the jet height at x=0. Under the same velocity and the nozzle clearance, the turbulent kinetic energy of flat -bottom was the smallest, and the turbulent kinetic energy of hemispherical-bottom was the largest. Under the same velocity, the pressure of hemispherical-bottom was the largest, wherever the pressure of flat-bottom was the smallest. The results can be theoretical basis for the optimum design and scale-up of the jet mixer.", "fno": "05514003", "keywords": [ "Computational Fluid Dynamics", "Confined Flow", "Flow Simulation", "Jets", "Mixing", "Nozzles", "Numerical Analysis", "Tanks Containers", "Turbulence", "Numerical Simulation", "Flow Characteristics", "Radial Distributions", "Turbulent Kinetic Energy", "Geometrical Jet Flow", "CFD Method", "RNG K Ε Turbulent Model", "Turbulence Pressure", "Jet Nozzle", "Hemispherical Bottom Pressure", "Flat Bottom Pressure", "Jet Mixer", "Numerical Simulation", "Kinetic Energy", "Equations", "Computational Fluid Dynamics", "Distributed Computing", "Educational Institutions", "Mechanical Engineering", "Chemical Technology", "Solid Modeling", "Predictive Models", "Jet Mixer", "Turbulent Flow", "Mean Velocity", "Turbulence Kinetic Energy" ], "authors": [ { "affiliation": null, "fullName": "Meng Hui-bo", "givenName": "Meng", "surname": "Hui-bo", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yu Yan-fang", "givenName": "Yu", "surname": "Yan-fang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Wu Jian-hua", "givenName": "Wu", "surname": "Jian-hua", "__typename": "ArticleAuthorType" } ], "idPrefix": "icic", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-06-01T00:00:00", "pubType": "proceedings", "pages": "3-6", "year": "2010", "issn": "2160-7443", "isbn": "978-1-4244-7081-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05514131", "articleId": "12OmNBC8Axv", "__typename": "AdjacentArticleType" }, "next": { "fno": "05514004", "articleId": "12OmNxG1yXJ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cdciem/2011/4350/0/4350b258", "title": "Numerical Study of High-Temperature Air Combustion Using Different Jet Nozzle", "doi": null, "abstractUrl": "/proceedings-article/cdciem/2011/4350b258/12OmNBCqbF3", "parentPublication": { "id": "proceedings/cdciem/2011/4350/0", "title": "Computer Distributed Control and Intelligent Environmental Monitoring, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hpcmp-ugc/2010/986/0/06017991", "title": "Numerical Modeling of Turbulent, Parallel, Round Jets", "doi": null, "abstractUrl": "/proceedings-article/hpcmp-ugc/2010/06017991/12OmNBSSVd3", "parentPublication": { "id": "proceedings/hpcmp-ugc/2010/986/0", "title": "2010 DoD High Performance Computing Modernization Program Users Group Conference (HPCMP-UGC 2010)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icic/2011/688/0/05954558", "title": "Numerical Study of the Turbulent Characteristic in a Novel Circulating Jet Mixer", "doi": null, "abstractUrl": "/proceedings-article/icic/2011/05954558/12OmNwDj17J", "parentPublication": { "id": "proceedings/icic/2011/688/0", "title": "2011 Fourth International Conference on Information and Computing (ICIC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccsee/2012/4647/2/4647b567", "title": "Numerical Study of Turbulent Mixing Processes in RQL Gas Turbine Combustor", "doi": null, "abstractUrl": "/proceedings-article/iccsee/2012/4647b567/12OmNx6PiAK", "parentPublication": { "id": null, "title": null, "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icic/2011/688/0/05954569", "title": "A Numerical Investigation of Effects of Different Nozzle Spacing on Turbulent Flow in a Novel Circular Jet Mixer", "doi": null, "abstractUrl": "/proceedings-article/icic/2011/05954569/12OmNxXUhO9", "parentPublication": { "id": "proceedings/icic/2011/688/0", "title": "2011 Fourth International Conference on Information and Computing (ICIC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icic/2009/3634/4/3634d358", "title": "Study on Numerical Simulation of Single-Phase Injection Device Flow Flied", "doi": null, "abstractUrl": "/proceedings-article/icic/2009/3634d358/12OmNy4IEY3", "parentPublication": { "id": "proceedings/icic/2009/3634/4", "title": "2009 Second International Conference on Information and Computing Science", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cdciem/2011/4350/0/4350b253", "title": "Numerical Study of Effect of Burner Jet Parameters on High Temperature Air Combustion of Coal Gas", "doi": null, "abstractUrl": "/proceedings-article/cdciem/2011/4350b253/12OmNyRxFBh", "parentPublication": { "id": "proceedings/cdciem/2011/4350/0", "title": "Computer Distributed Control and Intelligent Environmental Monitoring, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/itme/2015/8302/0/8302a506", "title": "Numerical Analysis Study on Flow Field and Sound Field in Spiral Tube", "doi": null, "abstractUrl": "/proceedings-article/itme/2015/8302a506/12OmNzT7Osa", "parentPublication": { "id": "proceedings/itme/2015/8302/0", "title": "2015 7th International Conference on Information Technology in Medicine and Education (ITME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmtma/2009/3583/2/3583b271", "title": "Numerical Simulation of Jet Flow Led by Different Central Body Locations", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2009/3583b271/12OmNzV70Dv", "parentPublication": { "id": "proceedings/icmtma/2009/3583/2", "title": "2009 International Conference on Measuring Technology and Mechatronics Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wcmeim/2019/5045/0/504500a401", "title": "A Comparative Study on Different Methods of Measuring Dynamic Characteristics of the Plasma Synthetic Jet", "doi": null, "abstractUrl": "/proceedings-article/wcmeim/2019/504500a401/1hHLuFzV4je", "parentPublication": { "id": "proceedings/wcmeim/2019/5045/0", "title": "2019 2nd World Conference on Mechanical Engineering and Intelligent Manufacturing (WCMEIM)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxE2mVE", "title": "Energy and Environment Technology, International Conference on", "acronym": "iceet", "groupId": "1003015", "volume": "2", "displayVolume": "2", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNxZBSC3", "doi": "10.1109/ICEET.2009.386", "title": "Kinetic Modeling of MTBE Degradation by Stabilized Immobilized Beads", "normalizedTitle": "Kinetic Modeling of MTBE Degradation by Stabilized Immobilized Beads", "abstract": "The contamination of ground and surface water with methyl tert-butyl ether (MTBE) has evoked substantial attention due to the frequent occurrence of storage tank leakage. Kinetic analysis of MTBE degradation by polyethyleneimine stabilized beads with alginate immobilized Methylibium petroleiphilum PM1 was investigated in this paper. The results showed that the stabilized beads could be used wider in respect to pH value, temperature and initial MTBE concentration than freely-suspended cells under the same conditions. The biochemical reaction rather than intraparticle diffusion resistance was considered as the key step by the kinetic analysis.", "abstracts": [ { "abstractType": "Regular", "content": "The contamination of ground and surface water with methyl tert-butyl ether (MTBE) has evoked substantial attention due to the frequent occurrence of storage tank leakage. Kinetic analysis of MTBE degradation by polyethyleneimine stabilized beads with alginate immobilized Methylibium petroleiphilum PM1 was investigated in this paper. The results showed that the stabilized beads could be used wider in respect to pH value, temperature and initial MTBE concentration than freely-suspended cells under the same conditions. The biochemical reaction rather than intraparticle diffusion resistance was considered as the key step by the kinetic analysis.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The contamination of ground and surface water with methyl tert-butyl ether (MTBE) has evoked substantial attention due to the frequent occurrence of storage tank leakage. Kinetic analysis of MTBE degradation by polyethyleneimine stabilized beads with alginate immobilized Methylibium petroleiphilum PM1 was investigated in this paper. The results showed that the stabilized beads could be used wider in respect to pH value, temperature and initial MTBE concentration than freely-suspended cells under the same conditions. The biochemical reaction rather than intraparticle diffusion resistance was considered as the key step by the kinetic analysis.", "fno": "3819b613", "keywords": [ "MTBE", "Methylibium Petroleiphilum PM 1", "Biodegradation", "Stabilized Immobilized Cells", "Kinetic" ], "authors": [ { "affiliation": null, "fullName": "Dongzhi Chen", "givenName": "Dongzhi", "surname": "Chen", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zhuowei Cheng", "givenName": "Zhuowei", "surname": "Cheng", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jianmeng Chen", "givenName": "Jianmeng", "surname": "Chen", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jing Chen", "givenName": "Jing", "surname": "Chen", "__typename": "ArticleAuthorType" } ], "idPrefix": "iceet", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-10-01T00:00:00", "pubType": "proceedings", "pages": "613-615", "year": "2009", "issn": null, "isbn": "978-0-7695-3819-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3819b609", "articleId": "12OmNxwENkD", "__typename": "AdjacentArticleType" }, "next": { "fno": "3819b616", "articleId": "12OmNzwpUlQ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/sc/2008/2835/0/05222734", "title": "0.374 Pflop/s trillion-particle kinetic modeling of laser plasma interaction on roadrunner", "doi": null, "abstractUrl": "/proceedings-article/sc/2008/05222734/12OmNARiM48", "parentPublication": { "id": "proceedings/sc/2008/2835/0", "title": "SC Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iri/2013/1050/0/06642478", "title": "Radiation therapy simulation and optimization using kinetic polygon modeling", "doi": null, "abstractUrl": "/proceedings-article/iri/2013/06642478/12OmNxEjXQx", "parentPublication": { "id": "proceedings/iri/2013/1050/0", "title": "2013 IEEE 14th International Conference on Information Reuse & Integration (IRI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceet/2009/3819/2/3819b492", "title": "Degradation of Azo Dye in Water through a Heterogeneous Fenton Process Catalyzed by Fe Alginate Gel Beads", "doi": null, "abstractUrl": "/proceedings-article/iceet/2009/3819b492/12OmNxw5Bl1", "parentPublication": { "id": "proceedings/iceet/2009/3819/2", "title": "Energy and Environment Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bibm/2007/3031/0/30310339", "title": "Kinetic Modeling Using BioPAX Ontology", "doi": null, "abstractUrl": "/proceedings-article/bibm/2007/30310339/12OmNy2agXP", "parentPublication": { "id": "proceedings/bibm/2007/3031/0", "title": "2007 IEEE International Conference on Bioinformatics and Biomedicine (BIBM 2007)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmtma/2010/3962/2/3962c558", "title": "Kinematics Modeling for Satellite Antenna Dish Stabilized Platform", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2010/3962c558/12OmNz2kqgZ", "parentPublication": { "id": "proceedings/icmtma/2010/3962/2", "title": "2010 International Conference on Measuring Technology and Mechatronics Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tb/2012/01/ttb2012010040", "title": "An Efficient Method for Modeling Kinetic Behavior of Channel Proteins in Cardiomyocytes", "doi": null, "abstractUrl": "/journal/tb/2012/01/ttb2012010040/13rRUyXKxSW", "parentPublication": { "id": "trans/tb", "title": "IEEE/ACM Transactions on Computational Biology and Bioinformatics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzSh1aC", "title": "2012 Third International Conference on Digital Manufacturing & Automation", "acronym": "icdma", "groupId": "1800272", "volume": "0", "displayVolume": "0", "year": "2012", "__typename": "ProceedingType" }, "article": { "id": "12OmNyQ7FDv", "doi": "10.1109/ICDMA.2012.148", "title": "Progress in Numerical Simulation of High Entrained Air-Water Two-Phase Flow", "normalizedTitle": "Progress in Numerical Simulation of High Entrained Air-Water Two-Phase Flow", "abstract": "The progress of numerical simulation for high entrained gas-liquid two phase flow were summarized and analyzed. The results show that the Euler-Euler method is the most suitable for the air-water two-phase flow. The advantages and disadvantages of three multiphase flow models (VOF, Mixture and Eulerian model) are analyzed and the results show that the Eulerian model is most popular and accurate for simulation the strong turbulent air-water two-phase flow, especially needing to consider the conversion of continuous and discrete phase. The coupled Eulerian model and mixture turbulence model is the best choice for simulation the flow field on the strong air entrained flow in hydraulic engineering.", "abstracts": [ { "abstractType": "Regular", "content": "The progress of numerical simulation for high entrained gas-liquid two phase flow were summarized and analyzed. The results show that the Euler-Euler method is the most suitable for the air-water two-phase flow. The advantages and disadvantages of three multiphase flow models (VOF, Mixture and Eulerian model) are analyzed and the results show that the Eulerian model is most popular and accurate for simulation the strong turbulent air-water two-phase flow, especially needing to consider the conversion of continuous and discrete phase. The coupled Eulerian model and mixture turbulence model is the best choice for simulation the flow field on the strong air entrained flow in hydraulic engineering.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The progress of numerical simulation for high entrained gas-liquid two phase flow were summarized and analyzed. The results show that the Euler-Euler method is the most suitable for the air-water two-phase flow. The advantages and disadvantages of three multiphase flow models (VOF, Mixture and Eulerian model) are analyzed and the results show that the Eulerian model is most popular and accurate for simulation the strong turbulent air-water two-phase flow, especially needing to consider the conversion of continuous and discrete phase. The coupled Eulerian model and mixture turbulence model is the best choice for simulation the flow field on the strong air entrained flow in hydraulic engineering.", "fno": "4772a626", "keywords": [ "Mathematical Model", "Atmospheric Modeling", "Computational Modeling", "Numerical Models", "Solid Modeling", "Equations", "Numerical Simulation", "Turbulence Model", "Numerical Simulation", "Air Water Two Phase Flow", "Multiphase Model" ], "authors": [ { "affiliation": null, "fullName": "Cheng Xiangju", "givenName": "Cheng", "surname": "Xiangju", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Chen Xuewei", "givenName": "Chen", "surname": "Xuewei", "__typename": "ArticleAuthorType" } ], "idPrefix": "icdma", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2012-07-01T00:00:00", "pubType": "proceedings", "pages": "626-629", "year": "2012", "issn": null, "isbn": "978-1-4673-2217-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4772a622", "articleId": "12OmNBC8Axg", "__typename": "AdjacentArticleType" }, "next": { "fno": "4772a630", "articleId": "12OmNCbkQCM", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/mcsul/2009/3976/0/3976a049", "title": "Evaluation of Air -- Water Flow in an Evaporative Condenser", "doi": null, "abstractUrl": "/proceedings-article/mcsul/2009/3976a049/12OmNANTArR", "parentPublication": { "id": "proceedings/mcsul/2009/3976/0", "title": "Computational Modeling, Southern Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceet/2009/3819/3/3819c423", "title": "Numerical Simulation on Smoke of Kitchen in an Apartment Unit", "doi": null, "abstractUrl": "/proceedings-article/iceet/2009/3819c423/12OmNAWpys2", "parentPublication": { "id": "proceedings/iceet/2009/3819/3", "title": "Energy and Environment Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceet/2009/3819/1/3819a218", "title": "Numerical Simulation of Gas/Solid Flow in a Novel Annular Spouted Bed with Multiple Gas Nozzles", "doi": null, "abstractUrl": "/proceedings-article/iceet/2009/3819a218/12OmNCesr6k", "parentPublication": { "id": "proceedings/iceet/2009/3819/1", "title": "Energy and Environment Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sc/2012/0806/0/1000a039", "title": "High throughput software for direct numerical simulations of compressible two-phase flows", "doi": null, "abstractUrl": "/proceedings-article/sc/2012/1000a039/12OmNrMHOl1", "parentPublication": { "id": "proceedings/sc/2012/0806/0", "title": "SC Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iceet/2009/3819/3/3819c207", "title": "Applying Numerical Simulation to Analyse the Performance of Nozzles", "doi": null, "abstractUrl": "/proceedings-article/iceet/2009/3819c207/12OmNwErpMB", "parentPublication": { "id": "proceedings/iceet/2009/3819/3", "title": "Energy and Environment Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscsct/2008/3498/1/3498a499", "title": "Numerical Simulation of Water-Filling Process in Long-Distance Diversion Tunnel with Free Flow", "doi": null, "abstractUrl": "/proceedings-article/iscsct/2008/3498a499/12OmNxWuiml", "parentPublication": { "id": null, "title": null, "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/esiat/2009/3682/3/3682c637", "title": "Analysis of Flow Field in Vertical Roller Mill Chamber Based on Turbulent Models", "doi": null, "abstractUrl": "/proceedings-article/esiat/2009/3682c637/12OmNz2TCCI", "parentPublication": { "id": "proceedings/esiat/2009/3682/3", "title": "Environmental Science and Information Application Technology, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icisce/2017/3013/0/3013b263", "title": "Study on Temperature Control Method of High Speed Ball Bearing with Oil-Air Lubrication", "doi": null, "abstractUrl": "/proceedings-article/icisce/2017/3013b263/12OmNzZWbN2", "parentPublication": { "id": "proceedings/icisce/2017/3013/0", "title": "2017 4th International Conference on Information Science and Control Engineering (ICISCE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2019/01/08663522", "title": "Visualization of Clouds and Atmospheric Air Flows", "doi": null, "abstractUrl": "/magazine/cg/2019/01/08663522/18exz8mOsPC", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/macise/2020/6695/0/09195600", "title": "Condensation of Humid Air Predicted by Numerical Model and Compared With Experiment", "doi": null, "abstractUrl": "/proceedings-article/macise/2020/09195600/1n7nIsWBglO", "parentPublication": { "id": "proceedings/macise/2020/6695/0", "title": "2020 International Conference on Mathematics and Computers in Science and Engineering (MACISE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvRU0lf", "title": "HPCMP Users Group Conference", "acronym": "hpcmp-ugc", "groupId": "1002962", "volume": "0", "displayVolume": "0", "year": "2009", "__typename": "ProceedingType" }, "article": { "id": "12OmNzdoME1", "doi": "10.1109/HPCMP-UGC.2009.21", "title": "Simulation of Mach 3 Cylinder Flow Using Kinetic and Continuum Solvers", "normalizedTitle": "Simulation of Mach 3 Cylinder Flow Using Kinetic and Continuum Solvers", "abstract": "The objective of this work is to study the performance of the unified kinetic/continuum solver referred to as the Unified Flow Solver (UFS) for a Mach 3 flow past a cylinder by comparing its results from those of a traditional Navier-Stokes equation solver. The intention is to benchmark and validate UFS to appeal to a wider group of users interested in solving flow problems of practical applications in the kinetic-continuum flight regime. This unified computational tool is being developed under the sponsorship of the Air Force Research Laboratory (AFRL) to solve both rarefied and continuum flow regimes. Some of the problems where such a solver would be used are re-entry vehicles, exhaust nozzle and plume flows, and MEMS/NANO devices, where a diverse range of conditions from continuum, to transition and rarefied flow regimes are encountered.", "abstracts": [ { "abstractType": "Regular", "content": "The objective of this work is to study the performance of the unified kinetic/continuum solver referred to as the Unified Flow Solver (UFS) for a Mach 3 flow past a cylinder by comparing its results from those of a traditional Navier-Stokes equation solver. The intention is to benchmark and validate UFS to appeal to a wider group of users interested in solving flow problems of practical applications in the kinetic-continuum flight regime. This unified computational tool is being developed under the sponsorship of the Air Force Research Laboratory (AFRL) to solve both rarefied and continuum flow regimes. Some of the problems where such a solver would be used are re-entry vehicles, exhaust nozzle and plume flows, and MEMS/NANO devices, where a diverse range of conditions from continuum, to transition and rarefied flow regimes are encountered.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The objective of this work is to study the performance of the unified kinetic/continuum solver referred to as the Unified Flow Solver (UFS) for a Mach 3 flow past a cylinder by comparing its results from those of a traditional Navier-Stokes equation solver. The intention is to benchmark and validate UFS to appeal to a wider group of users interested in solving flow problems of practical applications in the kinetic-continuum flight regime. This unified computational tool is being developed under the sponsorship of the Air Force Research Laboratory (AFRL) to solve both rarefied and continuum flow regimes. Some of the problems where such a solver would be used are re-entry vehicles, exhaust nozzle and plume flows, and MEMS/NANO devices, where a diverse range of conditions from continuum, to transition and rarefied flow regimes are encountered.", "fno": "3946a114", "keywords": [ "Supersonic Flow", "Confined Flow", "External Flows", "Flow Simulation", "Knudsen Flow", "Mach Number", "Micromechanical Devices", "Nozzles", "Knudsen Number", "Mach 3 Cylinder Flow Simulation", "Unified Kinetic Solver", "Unified Continuum Solver", "Navier Stokes Equation", "Kinetic Continuum Flight Regime", "Unified Computational Tool", "Air Force Research Laboratory", "Continuum Flow Regime", "Re Entry Vehicles", "Exhaust Nozzle", "NANO Device", "MEMS Device", "Rarefied Flow Regime", "Transition Flow Regime", "Kinetic Theory", "Computational Fluid Dynamics", "Heat Transfer", "Computational Modeling", "Stress", "Boundary Conditions", "Numerical Models" ], "authors": [], "idPrefix": "hpcmp-ugc", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2009-06-01T00:00:00", "pubType": "proceedings", "pages": "114-118", "year": "2009", "issn": null, "isbn": "978-0-7695-3946-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "3946a106", "articleId": "12OmNrFkeRW", "__typename": "AdjacentArticleType" }, "next": { "fno": "3946a119", "articleId": "12OmNAXPyly", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/sc/1998/8707/0/87070015", "title": "Agent Middleware for Heterogeneous Scientific Simulations", "doi": null, "abstractUrl": "/proceedings-article/sc/1998/87070015/12OmNvwC5wa", "parentPublication": { "id": "proceedings/sc/1998/8707/0", "title": "SC Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2015/9403/0/9403a193", "title": "A GUI for Urban Wind Flow CFD Analysis of Small Scale Wind Applications", "doi": null, "abstractUrl": "/proceedings-article/cw/2015/9403a193/12OmNyGtjdR", "parentPublication": { "id": "proceedings/cw/2015/9403/0", "title": "2015 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/rtcsa/2012/4824/0/4824a232", "title": "Performance Comparisons of Parallel Power Flow Solvers on GPU System", "doi": null, "abstractUrl": "/proceedings-article/rtcsa/2012/4824a232/12OmNyoSb8t", "parentPublication": { "id": "proceedings/rtcsa/2012/4824/0", "title": "2012 IEEE International Conference on Embedded and Real-Time Computing Systems and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sc/1989/341/0/05349029", "title": "Computational aerothermodynamics", "doi": null, "abstractUrl": "/proceedings-article/sc/1989/05349029/12OmNzzP5O8", "parentPublication": { "id": "proceedings/sc/1989/341/0", "title": "Proceedings of the 1989 ACM/IEEE Conference on Supercomputing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/1994/11/i1133", "title": "On Poisson Solvers and Semi-Direct Methods for Computing Area Based Optical Flow", "doi": null, "abstractUrl": "/journal/tp/1994/11/i1133/13rRUxC0SX5", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08986688", "title": "Kinetic-Based Multiphase Flow Simulation", "doi": null, "abstractUrl": "/journal/tg/2021/07/08986688/1hed9kswQBW", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/acomp/2019/4723/0/472300a103", "title": "Path Conservative WENO Schemes and Riemann Solvers for Continuum Mechanics", "doi": null, "abstractUrl": "/proceedings-article/acomp/2019/472300a103/1ivu52xNvFe", "parentPublication": { "id": "proceedings/acomp/2019/4723/0", "title": "2019 International Conference on Advanced Computing and Applications (ACOMP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "17D45VtKiqs", "title": "2018 International Conference on Virtual Reality and Intelligent Systems (ICVRIS)", "acronym": "icvris", "groupId": "1828444", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "17D45WaTkjb", "doi": "10.1109/ICVRIS.2018.00026", "title": "Numerical Simulation of Three Dimensional Flow Field in Water Treatment Agitator Based on Fluent", "normalizedTitle": "Numerical Simulation of Three Dimensional Flow Field in Water Treatment Agitator Based on Fluent", "abstract": "Because of the complex flow phenomenon in the mixing equipment, the design of the mixing equipment depends on the actual operation experience of the engineers. With the increasing application of numerical simulation technology in various fields, the application of numerical simulation technology to the design of mixing equipment has a great advantage over the traditional experience design method. In this paper, the three-dimensional flow field of the water treatment agitator is simulated by two modeling methods, and the characteristics of the flow field are compared and analyzed, which will provide some reference for the optimization design of the water treatment process.", "abstracts": [ { "abstractType": "Regular", "content": "Because of the complex flow phenomenon in the mixing equipment, the design of the mixing equipment depends on the actual operation experience of the engineers. With the increasing application of numerical simulation technology in various fields, the application of numerical simulation technology to the design of mixing equipment has a great advantage over the traditional experience design method. In this paper, the three-dimensional flow field of the water treatment agitator is simulated by two modeling methods, and the characteristics of the flow field are compared and analyzed, which will provide some reference for the optimization design of the water treatment process.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Because of the complex flow phenomenon in the mixing equipment, the design of the mixing equipment depends on the actual operation experience of the engineers. With the increasing application of numerical simulation technology in various fields, the application of numerical simulation technology to the design of mixing equipment has a great advantage over the traditional experience design method. In this paper, the three-dimensional flow field of the water treatment agitator is simulated by two modeling methods, and the characteristics of the flow field are compared and analyzed, which will provide some reference for the optimization design of the water treatment process.", "fno": "803100a075", "keywords": [ "Computational Fluid Dynamics", "Design Engineering", "Mixing", "Numerical Analysis", "Production Equipment", "Water Treatment", "Water Treatment Agitator", "Numerical Simulation Technology", "Three Dimensional Flow Field", "Mixing Equipment Design", "Mathematical Model", "Computational Modeling", "Numerical Models", "Kinetic Energy", "Numerical Simulation", "Shafts", "Blades", "Agitator", "Coagulation Sedimentation Process", "Numerical Simulation" ], "authors": [ { "affiliation": null, "fullName": "Runjian Dong", "givenName": "Runjian", "surname": "Dong", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Wenfu Wu", "givenName": "Wenfu", "surname": "Wu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Feng Liu", "givenName": "Feng", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Xinyang Wang", "givenName": "Xinyang", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Liang Zhang", "givenName": "Liang", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Junxing Li", "givenName": "Junxing", "surname": "Li", "__typename": "ArticleAuthorType" } ], "idPrefix": "icvris", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-08-01T00:00:00", "pubType": "proceedings", "pages": "75-78", "year": "2018", "issn": null, "isbn": "978-1-5386-8031-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "803100a071", "articleId": "17D45XtvpeB", "__typename": "AdjacentArticleType" }, "next": { "fno": "803100a079", "articleId": "17D45WaTkoe", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icdma/2010/4286/1/4286a670", "title": "Evaluation of Micro-pore Ceramic Filtration and UV Radiation Combination on Ballast Water Treatment", "doi": null, "abstractUrl": "/proceedings-article/icdma/2010/4286a670/12OmNCdk2D9", "parentPublication": { "id": "proceedings/icdma/2010/4286/1", "title": "2010 International Conference on Digital Manufacturing & Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icisce/2015/6850/0/6850a670", "title": "Comprehensive Analysis on Calculation Iterative Methods for Natural River and Open Channel Water Curve", "doi": null, "abstractUrl": "/proceedings-article/icisce/2015/6850a670/12OmNqIzh51", "parentPublication": { "id": "proceedings/icisce/2015/6850/0", "title": "2015 2nd International Conference on Information Science and Control Engineering (ICISCE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icic/2010/4047/3/4047c292", "title": "Three-Dimension Numerical Simulation of Viscosity Field in the Rubber Mixing Process", "doi": null, "abstractUrl": "/proceedings-article/icic/2010/4047c292/12OmNrMZpB9", "parentPublication": { "id": "proceedings/icic/2010/4047/3", "title": "2010 Third International Conference on Information and Computing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmcce/2017/2628/0/2628a220", "title": "Numerical Simulation Method for Three-Dimensional Flow Field of Multistage Axial Flow Compressor", "doi": null, "abstractUrl": "/proceedings-article/icmcce/2017/2628a220/12OmNrYCY0n", "parentPublication": { "id": "proceedings/icmcce/2017/2628/0", "title": "2017 Second International Conference on Mechanical, Control and Computer Engineering (ICMCCE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccis/2011/4501/0/4501a549", "title": "The Simulation of Non-heating Pipelines' Temperature Drop under Stopping Mixing Water Condition", "doi": null, "abstractUrl": "/proceedings-article/iccis/2011/4501a549/12OmNwudQN5", "parentPublication": { "id": "proceedings/iccis/2011/4501/0", "title": "2011 International Conference on Computational and Information Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmtma/2013/4932/0/4932a737", "title": "Numerical Simulation and Analysis of Water-Cooled Radiator Performance for Traction Converter", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2013/4932a737/12OmNx7ov09", "parentPublication": { "id": "proceedings/icmtma/2013/4932/0", "title": "2013 Fifth International Conference on Measuring Technology and Mechatronics Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icrccs/2009/3927/0/3927a085", "title": "Three-Dimensional Flow Numerical Simulation of the Special Emulsifying Device Emulsifying Process", "doi": null, "abstractUrl": "/proceedings-article/icrccs/2009/3927a085/12OmNxHrygk", "parentPublication": { "id": "proceedings/icrccs/2009/3927/0", "title": "Research Challenges in Computer Science, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdma/2012/4772/0/4772a626", "title": "Progress in Numerical Simulation of High Entrained Air-Water Two-Phase Flow", "doi": null, "abstractUrl": "/proceedings-article/icdma/2012/4772a626/12OmNyQ7FDv", "parentPublication": { "id": "proceedings/icdma/2012/4772/0", "title": "2012 Third International Conference on Digital Manufacturing & Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccis/2010/4270/0/4270a089", "title": "Numerical Simulation of the External Flow-Field of Pulsed Anti-riots Water-Cannon", "doi": null, "abstractUrl": "/proceedings-article/iccis/2010/4270a089/12OmNz5JBUf", "parentPublication": { "id": "proceedings/iccis/2010/4270/0", "title": "2010 International Conference on Computational and Information Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/aiea/2020/8288/0/828800a659", "title": "Research on Reasonable Parameter Matching of Vertical Vibration Stirring Device Based on EDEM", "doi": null, "abstractUrl": "/proceedings-article/aiea/2020/828800a659/1nTuq5yOjhC", "parentPublication": { "id": "proceedings/aiea/2020/8288/0", "title": "2020 International Conference on Artificial Intelligence and Electromechanical Automation (AIEA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1jIxhEnA8IE", "title": "2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)", "acronym": "vrw", "groupId": "1836626", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1jIxpVaNpEQ", "doi": "10.1109/VRW50115.2020.00179", "title": "Robust turbulence simulation for particle-based fluids using the Rankine vortex model", "normalizedTitle": "Robust turbulence simulation for particle-based fluids using the Rankine vortex model", "abstract": "We propose a novel turbulence refinement method based on the Rankine vortex model for SPH (smoothed particle hydrodynamics) simulations. Surface details are enhanced by recovering the energy lost in the rotational degrees of freedom of SPH particles. The Rankine vortex model is used to convert the diffused and stretched angular kinetic energy of particles to the linear kinetic energy of their neighbours. Our model naturally prevents the positive feedback effect between the velocity and vorticity fields since the vortex model is designed to alter the velocity without introducing external sources. Experimental results show that our method can recover missing high-frequency details realistically and maintain convergence in both static and highly dynamic scenarios.", "abstracts": [ { "abstractType": "Regular", "content": "We propose a novel turbulence refinement method based on the Rankine vortex model for SPH (smoothed particle hydrodynamics) simulations. Surface details are enhanced by recovering the energy lost in the rotational degrees of freedom of SPH particles. The Rankine vortex model is used to convert the diffused and stretched angular kinetic energy of particles to the linear kinetic energy of their neighbours. Our model naturally prevents the positive feedback effect between the velocity and vorticity fields since the vortex model is designed to alter the velocity without introducing external sources. Experimental results show that our method can recover missing high-frequency details realistically and maintain convergence in both static and highly dynamic scenarios.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose a novel turbulence refinement method based on the Rankine vortex model for SPH (smoothed particle hydrodynamics) simulations. Surface details are enhanced by recovering the energy lost in the rotational degrees of freedom of SPH particles. The Rankine vortex model is used to convert the diffused and stretched angular kinetic energy of particles to the linear kinetic energy of their neighbours. Our model naturally prevents the positive feedback effect between the velocity and vorticity fields since the vortex model is designed to alter the velocity without introducing external sources. Experimental results show that our method can recover missing high-frequency details realistically and maintain convergence in both static and highly dynamic scenarios.", "fno": "09090460", "keywords": [ "Computational Fluid Dynamics", "Hydrodynamics", "Smoothed Particle Hydrodynamics", "Turbulence", "Vortices", "Particle Based Fluids", "Rankine Vortex Model", "Particle Hydrodynamics", "SPH Particles", "Angular Kinetic Energy", "Linear Kinetic Energy", "Robust Turbulence Simulation", "Turbulence Refinement Method", "Vorticity Fields", "Solid Modeling", "Computational Modeling", "Graphics", "Numerical Models", "Kinetic Energy", "Angular Velocity", "Indexes", "Computing Methodologies", "Computer Graphics", "Animation", "Physical Simulation" ], "authors": [ { "affiliation": "University of Science and Technology,Beijing Advanced Innovation Center for Materials Genome Engineering,Beijing", "fullName": "Xiaokun Wang", "givenName": "Xiaokun", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Science and Technology,Beijing Advanced Innovation Center for Materials Genome Engineering,Beijing", "fullName": "Sinuo Liu", "givenName": "Sinuo", "surname": "Liu", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Science and Technology,Beijing Advanced Innovation Center for Materials Genome Engineering,Beijing", "fullName": "Xiaojuan Ban", "givenName": "Xiaojuan", "surname": "Ban", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Science and Technology,Beijing Advanced Innovation Center for Materials Genome Engineering,Beijing", "fullName": "Yanrui Xu", "givenName": "Yanrui", "surname": "Xu", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Science and Technology,Beijing Advanced Innovation Center for Materials Genome Engineering,Beijing", "fullName": "Jing Zhou", "givenName": "Jing", "surname": "Zhou", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Groningen,Bernoulli Institute", "fullName": "Jiří Kosinka", "givenName": "Jiří", "surname": "Kosinka", "__typename": "ArticleAuthorType" } ], "idPrefix": "vrw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-03-01T00:00:00", "pubType": "proceedings", "pages": "656-657", "year": "2020", "issn": null, "isbn": "978-1-7281-6532-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09090401", "articleId": "1jIxmhXvH7a", "__typename": "AdjacentArticleType" }, "next": { "fno": "09090429", "articleId": "1jIxznjCduE", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/hpcmp-ugc/2010/986/0/06018013", "title": "Unitary Quantum Lattice Gas Algorithms for Quantum to Classical Turbulence", "doi": null, "abstractUrl": "/proceedings-article/hpcmp-ugc/2010/06018013/12OmNwswg0I", "parentPublication": { "id": "proceedings/hpcmp-ugc/2010/986/0", "title": "2010 DoD High Performance Computing Modernization Program Users Group Conference (HPCMP-UGC 2010)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/paciia/2008/3490/2/3490c634", "title": "Smoothed Particle Hydrodynamics for Numerical Simulation of Duct Conveying", "doi": null, "abstractUrl": "/proceedings-article/paciia/2008/3490c634/12OmNyQ7FEy", "parentPublication": { "id": "paciia/2008/3490/2", "title": "Pacific-Asia Workshop on Computational Intelligence and Industrial Application, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cad-graphics/2013/2576/0/06815003", "title": "Synthesizing Solid-Induced Turbulence for Particle-Based Fluids", "doi": null, "abstractUrl": "/proceedings-article/cad-graphics/2013/06815003/12OmNzlD9EF", "parentPublication": { "id": "proceedings/cad-graphics/2013/2576/0", "title": "2013 International Conference on Computer-Aided Design and Computer Graphics (CAD/Graphics)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2014/4258/0/4258a065", "title": "SPH Fluids for Viscous Jet Buckling", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2014/4258a065/12OmNzn38LQ", "parentPublication": { "id": "proceedings/sibgrapi/2014/4258/0", "title": "2014 27th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/03/07487018", "title": "Divergence-Free SPH for Incompressible and Viscous Fluids", "doi": null, "abstractUrl": "/journal/tg/2017/03/07487018/13rRUxASuhF", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2015/06/mcg2015060060", "title": "Vortex Filaments in Grids for Scalable, Fine Smoke Simulation", "doi": null, "abstractUrl": "/magazine/cg/2015/06/mcg2015060060/13rRUxASuk3", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/03/07845705", "title": "Animating Wall-Bounded Turbulent Smoke via Filament-Mesh Particle-Particle Method", "doi": null, "abstractUrl": "/journal/tg/2018/03/07845705/13rRUxBJhvD", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2019/06/08353115", "title": "Turbulent Micropolar SPH Fluids with Foam", "doi": null, "abstractUrl": "/journal/tg/2019/06/08353115/13rRUxNW1Zu", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2019/01/08440096", "title": "Objective Vortex Corelines of Finite-sized Objects in Fluid Flows", "doi": null, "abstractUrl": "/journal/tg/2019/01/08440096/17D45WXIkH8", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798224", "title": "Viscosity-based Vorticity Correction for Turbulent SPH Fluids", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798224/1cJ1aEpf4wo", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1BmEezmpGrm", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1BmFaDYHtgA", "doi": "10.1109/ICCV48922.2021.00119", "title": "CODEs: Chamfer Out-of-Distribution Examples against Overconfidence Issue", "normalizedTitle": "CODEs: Chamfer Out-of-Distribution Examples against Overconfidence Issue", "abstract": "Overconfident predictions on out-of-distribution (OOD) samples is a thorny issue for deep neural networks. The key to resolve the OOD overconfidence issue inherently is to build a subset of OOD samples and then suppress predictions on them. This paper proposes the Chamfer OOD examples (CODEs), whose distribution is close to that of in-distribution samples, and thus could be utilized to alleviate the OOD overconfidence issue effectively by suppressing predictions on them. To obtain CODEs, we first generate seed OOD examples via slicing&#x0026;splicing operations on in-distribution samples from different categories, and then feed them to the Chamfer generative adversarial network for distribution transformation, without accessing to any extra data. Training with suppressing predictions on CODEs is validated to alleviate the OOD overconfidence issue largely without hurting classification accuracy, and outperform the state-of-the-art methods. Besides, we demonstrate CODEs are useful for improving OOD detection and classification.", "abstracts": [ { "abstractType": "Regular", "content": "Overconfident predictions on out-of-distribution (OOD) samples is a thorny issue for deep neural networks. The key to resolve the OOD overconfidence issue inherently is to build a subset of OOD samples and then suppress predictions on them. This paper proposes the Chamfer OOD examples (CODEs), whose distribution is close to that of in-distribution samples, and thus could be utilized to alleviate the OOD overconfidence issue effectively by suppressing predictions on them. To obtain CODEs, we first generate seed OOD examples via slicing&#x0026;splicing operations on in-distribution samples from different categories, and then feed them to the Chamfer generative adversarial network for distribution transformation, without accessing to any extra data. Training with suppressing predictions on CODEs is validated to alleviate the OOD overconfidence issue largely without hurting classification accuracy, and outperform the state-of-the-art methods. Besides, we demonstrate CODEs are useful for improving OOD detection and classification.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Overconfident predictions on out-of-distribution (OOD) samples is a thorny issue for deep neural networks. The key to resolve the OOD overconfidence issue inherently is to build a subset of OOD samples and then suppress predictions on them. This paper proposes the Chamfer OOD examples (CODEs), whose distribution is close to that of in-distribution samples, and thus could be utilized to alleviate the OOD overconfidence issue effectively by suppressing predictions on them. To obtain CODEs, we first generate seed OOD examples via slicing&splicing operations on in-distribution samples from different categories, and then feed them to the Chamfer generative adversarial network for distribution transformation, without accessing to any extra data. Training with suppressing predictions on CODEs is validated to alleviate the OOD overconfidence issue largely without hurting classification accuracy, and outperform the state-of-the-art methods. Besides, we demonstrate CODEs are useful for improving OOD detection and classification.", "fno": "281200b133", "keywords": [ "Training", "Deep Learning", "Computer Vision", "Codes", "Neural Networks", "Training Data", "Generative Adversarial Networks" ], "authors": [ { "affiliation": "Guangzhou University", "fullName": "Keke Tang", "givenName": "Keke", "surname": "Tang", "__typename": "ArticleAuthorType" }, { "affiliation": "Guangzhou University", "fullName": "Dingruibo Miao", "givenName": "Dingruibo", "surname": "Miao", "__typename": "ArticleAuthorType" }, { "affiliation": "Guangzhou University", "fullName": "Weilong Peng", "givenName": "Weilong", "surname": "Peng", "__typename": "ArticleAuthorType" }, { "affiliation": "Guangzhou University", "fullName": "Jianpeng Wu", "givenName": "Jianpeng", "surname": "Wu", "__typename": "ArticleAuthorType" }, { "affiliation": "Guangzhou University", "fullName": "Yawen Shi", "givenName": "Yawen", "surname": "Shi", "__typename": "ArticleAuthorType" }, { "affiliation": "Guangzhou University", "fullName": "Zhaoquan Gu", "givenName": "Zhaoquan", "surname": "Gu", "__typename": "ArticleAuthorType" }, { "affiliation": "Guangzhou University", "fullName": "Zhihong Tian", "givenName": "Zhihong", "surname": "Tian", "__typename": "ArticleAuthorType" }, { "affiliation": "Texas A&M University", "fullName": "Wenping Wang", "givenName": "Wenping", "surname": "Wang", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-10-01T00:00:00", "pubType": "proceedings", "pages": "1133-1142", "year": "2021", "issn": null, "isbn": "978-1-6654-2812-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "281200b125", "articleId": "1BmKWnqkmxa", "__typename": "AdjacentArticleType" }, "next": { "fno": "281200b143", "articleId": "1BmKj5CIqek", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cogmi/2021/1621/0/162100a282", "title": "On Detection of Out of Distribution Inputs in Deep Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/cogmi/2021/162100a282/1CxzVIHwETu", "parentPublication": { "id": "proceedings/cogmi/2021/1621/0", "title": "2021 IEEE Third International Conference on Cognitive Machine Intelligence (CogMI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900c836", "title": "Class-wise Thresholding for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900c836/1G56Ht7f7jO", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900e350", "title": "PyTorch-OOD: A Library for Out-of-Distribution Detection based on PyTorch", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900e350/1G56ntxOSY0", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f520", "title": "Mixture Outlier Exposure: Towards Out-of-Distribution Detection in Fine-grained Environments", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f520/1KxUKOwyJJS", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600c602", "title": "Heatmap-based Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600c602/1L8qqu8Q3OU", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800k0948", "title": "Generalized ODIN: Detecting Out-of-Distribution Image Without Learning From Out-of-Distribution Data", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800k0948/1m3ofcCTYha", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2021/8808/0/09412489", "title": "NeuralFP: Out-of-distribution Detection using Fingerprints of Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icpr/2021/09412489/1tmhrKCdAFa", "parentPublication": { "id": "proceedings/icpr/2021/8808/0", "title": "2020 25th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2021/0191/0/019100d248", "title": "Uncertainty-aware GAN with Adaptive Loss for Robust MRI Image Enhancement", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2021/019100d248/1yNiDR4gRlS", "parentPublication": { "id": "proceedings/iccvw/2021/0191/0", "title": "2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2021/0191/0/019100d317", "title": "SOoD: Self-Supervised Out-of-Distribution Detection Under Domain Shift for Multi-Class Colorectal Cancer Tissue Types", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2021/019100d317/1yNiFB25sJi", "parentPublication": { "id": "proceedings/iccvw/2021/0191/0", "title": "2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2021/2354/0/235400a409", "title": "GCOOD: A Generic Coupled Out-of-Distribution Detector for Robust Classification", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2021/235400a409/1zurshXlLR6", "parentPublication": { "id": "proceedings/sibgrapi/2021/2354/0", "title": "2021 34th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1BmEezmpGrm", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1BmFq1tAb16", "doi": "10.1109/ICCV48922.2021.00819", "title": "Semantically Coherent Out-of-Distribution Detection", "normalizedTitle": "Semantically Coherent Out-of-Distribution Detection", "abstract": "Current out-of-distribution (OOD) detection benchmarks are commonly built by defining one dataset as in-distribution (ID) and all others as OOD. However, these benchmarks unfortunately introduce some unwanted and impractical goals, e.g., to perfectly distinguish CIFAR dogs from ImageNet dogs, even though they have the same semantics and negligible covariate shifts. These unrealistic goals will result in an extremely narrow range of model capabilities, greatly limiting their use in real applications. To overcome these drawbacks, we re-design the benchmarks and propose the semantically coherent out-of-distribution detection (SC-OOD). On the SC-OOD benchmarks, existing methods suffer from large performance degradation, suggesting that they are extremely sensitive to low-level discrepancy between data sources while ignoring their inherent semantics. To develop an effective SC-OOD detection approach, we leverage an external unlabeled set and design a concise framework featured by unsupervised dual grouping (UDG) for the joint modeling of ID and OOD data. The proposed UDG can not only enrich the semantic knowledge of the model by exploiting unlabeled data in an unsupervised manner, but also distinguish ID/OOD samples to enhance ID classification and OOD detection tasks simultaneously. Extensive experiments demonstrate that our approach achieves the state-of-the-art performance on SC-OOD benchmarks. Code and benchmarks are provided on our project page: https://jingkang50.github.io/projects/scood.", "abstracts": [ { "abstractType": "Regular", "content": "Current out-of-distribution (OOD) detection benchmarks are commonly built by defining one dataset as in-distribution (ID) and all others as OOD. However, these benchmarks unfortunately introduce some unwanted and impractical goals, e.g., to perfectly distinguish CIFAR dogs from ImageNet dogs, even though they have the same semantics and negligible covariate shifts. These unrealistic goals will result in an extremely narrow range of model capabilities, greatly limiting their use in real applications. To overcome these drawbacks, we re-design the benchmarks and propose the semantically coherent out-of-distribution detection (SC-OOD). On the SC-OOD benchmarks, existing methods suffer from large performance degradation, suggesting that they are extremely sensitive to low-level discrepancy between data sources while ignoring their inherent semantics. To develop an effective SC-OOD detection approach, we leverage an external unlabeled set and design a concise framework featured by unsupervised dual grouping (UDG) for the joint modeling of ID and OOD data. The proposed UDG can not only enrich the semantic knowledge of the model by exploiting unlabeled data in an unsupervised manner, but also distinguish ID/OOD samples to enhance ID classification and OOD detection tasks simultaneously. Extensive experiments demonstrate that our approach achieves the state-of-the-art performance on SC-OOD benchmarks. Code and benchmarks are provided on our project page: https://jingkang50.github.io/projects/scood.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Current out-of-distribution (OOD) detection benchmarks are commonly built by defining one dataset as in-distribution (ID) and all others as OOD. However, these benchmarks unfortunately introduce some unwanted and impractical goals, e.g., to perfectly distinguish CIFAR dogs from ImageNet dogs, even though they have the same semantics and negligible covariate shifts. These unrealistic goals will result in an extremely narrow range of model capabilities, greatly limiting their use in real applications. To overcome these drawbacks, we re-design the benchmarks and propose the semantically coherent out-of-distribution detection (SC-OOD). On the SC-OOD benchmarks, existing methods suffer from large performance degradation, suggesting that they are extremely sensitive to low-level discrepancy between data sources while ignoring their inherent semantics. To develop an effective SC-OOD detection approach, we leverage an external unlabeled set and design a concise framework featured by unsupervised dual grouping (UDG) for the joint modeling of ID and OOD data. The proposed UDG can not only enrich the semantic knowledge of the model by exploiting unlabeled data in an unsupervised manner, but also distinguish ID/OOD samples to enhance ID classification and OOD detection tasks simultaneously. Extensive experiments demonstrate that our approach achieves the state-of-the-art performance on SC-OOD benchmarks. Code and benchmarks are provided on our project page: https://jingkang50.github.io/projects/scood.", "fno": "281200i281", "keywords": [ "Degradation", "Limiting", "Soft Sensors", "Semantics", "Pipelines", "Dogs", "Benchmark Testing", "Transfer Low Shot Semi Unsupervised Learning", "Recognition And Classification" ], "authors": [ { "affiliation": "Nanyang Technological University,S-Lab", "fullName": "Jingkang Yang", "givenName": "Jingkang", "surname": "Yang", "__typename": "ArticleAuthorType" }, { "affiliation": "SenseTime Research", "fullName": "Haoqi Wang", "givenName": "Haoqi", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": "SenseTime Research", "fullName": "Litong Feng", "givenName": "Litong", "surname": "Feng", "__typename": "ArticleAuthorType" }, { "affiliation": "SenseTime Research", "fullName": "Xiaopeng Yan", "givenName": "Xiaopeng", "surname": "Yan", "__typename": "ArticleAuthorType" }, { "affiliation": "SenseTime Research", "fullName": "Huabin Zheng", "givenName": "Huabin", "surname": "Zheng", "__typename": "ArticleAuthorType" }, { "affiliation": "SenseTime Research", "fullName": "Wayne Zhang", "givenName": "Wayne", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": "Nanyang Technological University,S-Lab", "fullName": "Ziwei Liu", "givenName": "Ziwei", "surname": "Liu", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-10-01T00:00:00", "pubType": "proceedings", "pages": "8281-8289", "year": "2021", "issn": null, "isbn": "978-1-6654-2812-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1BmFpY2H1sI", "name": "piccv202128120-09711116s1-mm_281200i281.zip", "size": "435 kB", "location": "https://www.computer.org/csdl/api/v1/extra/piccv202128120-09711116s1-mm_281200i281.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "281200i271", "articleId": "1BmKKNFAMuI", "__typename": "AdjacentArticleType" }, "next": { "fno": "281200i290", "articleId": "1BmKPBl8fzq", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvprw/2022/8739/0/873900a163", "title": "RODD: A Self-Supervised Approach for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900a163/1G56R8MQAwg", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600e911", "title": "ViM: Out-Of-Distribution with Virtual-logit Matching", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600e911/1H0LoDIYyB2", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600t9195", "title": "Neural Mean Discrepancy for Efficient Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600t9195/1H1mfixU9JC", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f520", "title": "Mixture Outlier Exposure: Towards Out-of-Distribution Detection in Fine-grained Environments", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f520/1KxUKOwyJJS", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600c602", "title": "Heatmap-based Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600c602/1L8qqu8Q3OU", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300j517", "title": "Unsupervised Out-of-Distribution Detection by Maximum Classifier Discrepancy", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300j517/1hVl9mSRtfO", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08994105", "title": "OoDAnalyzer: Interactive Analysis of Out-of-Distribution Samples", "doi": null, "abstractUrl": "/journal/tg/2021/07/08994105/1hkQR7Os8SI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2021/8808/0/09412489", "title": "NeuralFP: Out-of-distribution Detection using Fingerprints of Neural Networks", "doi": null, "abstractUrl": "/proceedings-article/icpr/2021/09412489/1tmhrKCdAFa", "parentPublication": { "id": "proceedings/icpr/2021/8808/0", "title": "2020 25th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900p5308", "title": "MOOD: Multi-level Out-of-distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900p5308/1yeHIgRMMgg", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900j447", "title": "Out-of-Distribution Detection Using Union of 1-Dimensional Subspaces", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900j447/1yeKwz9Z2Uw", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1G55WEFExd6", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "acronym": "cvprw", "groupId": "1001809", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1G56R8MQAwg", "doi": "10.1109/CVPRW56347.2022.00028", "title": "RODD: A Self-Supervised Approach for Robust Out-of-Distribution Detection", "normalizedTitle": "RODD: A Self-Supervised Approach for Robust Out-of-Distribution Detection", "abstract": "Recent studies have started to address the concern of detecting and rejecting the out-of-distribution (OOD) samples as a major challenge in the safe deployment of deep learning (DL) models. It is desired that the DL model should only be confident about the in-distribution (ID) data which reinforces the driving principle of the OOD detection. In this paper, we propose a simple yet effective generalized OOD detection method independent of out-of-distribution datasets. Our approach relies on self-supervised feature learning of the training samples, where the embeddings lie on a compact low-dimensional space. Motivated by the recent studies that show self-supervised adversarial contrastive learning helps robustify the model, we empirically show that a pre-trained model with self-supervised contrastive learning yields a better model for uni-dimensional feature learning in the latent space. The method proposed in this work, referred to as RODD, outperforms SOTA detection performance on extensive suite of benchmark datasets on OOD detection tasks. On the CIFAR-100 benchmarks, RODD achieves a 26.97 % lower false positive rate (FPR@95) compared to SOTA methods. Our code is publicly available.<sup>1</sup>", "abstracts": [ { "abstractType": "Regular", "content": "Recent studies have started to address the concern of detecting and rejecting the out-of-distribution (OOD) samples as a major challenge in the safe deployment of deep learning (DL) models. It is desired that the DL model should only be confident about the in-distribution (ID) data which reinforces the driving principle of the OOD detection. In this paper, we propose a simple yet effective generalized OOD detection method independent of out-of-distribution datasets. Our approach relies on self-supervised feature learning of the training samples, where the embeddings lie on a compact low-dimensional space. Motivated by the recent studies that show self-supervised adversarial contrastive learning helps robustify the model, we empirically show that a pre-trained model with self-supervised contrastive learning yields a better model for uni-dimensional feature learning in the latent space. The method proposed in this work, referred to as RODD, outperforms SOTA detection performance on extensive suite of benchmark datasets on OOD detection tasks. On the CIFAR-100 benchmarks, RODD achieves a 26.97 % lower false positive rate (FPR@95) compared to SOTA methods. Our code is publicly available.<sup>1</sup>", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Recent studies have started to address the concern of detecting and rejecting the out-of-distribution (OOD) samples as a major challenge in the safe deployment of deep learning (DL) models. It is desired that the DL model should only be confident about the in-distribution (ID) data which reinforces the driving principle of the OOD detection. In this paper, we propose a simple yet effective generalized OOD detection method independent of out-of-distribution datasets. Our approach relies on self-supervised feature learning of the training samples, where the embeddings lie on a compact low-dimensional space. Motivated by the recent studies that show self-supervised adversarial contrastive learning helps robustify the model, we empirically show that a pre-trained model with self-supervised contrastive learning yields a better model for uni-dimensional feature learning in the latent space. The method proposed in this work, referred to as RODD, outperforms SOTA detection performance on extensive suite of benchmark datasets on OOD detection tasks. On the CIFAR-100 benchmarks, RODD achieves a 26.97 % lower false positive rate (FPR@95) compared to SOTA methods. Our code is publicly available.1", "fno": "873900a163", "keywords": [ "Deep Learning Artificial Intelligence", "Feature Extraction", "Object Detection", "Supervised Learning", "CIFAR 100 Benchmark", "Unidimensional Feature Learning", "Self Supervised Adversarial Contrastive Learning", "Generalized OOD Detection Method", "SOTA Detection Performance", "Latent Space", "Compact Low Dimensional Space", "Self Supervised Feature Learning", "In Distribution Data", "DL Model", "Deep Learning Models", "Robust Out Of Distribution Detection", "RODD", "Representation Learning", "Training", "Deep Learning", "Gaussian Noise", "Benchmark Testing", "Feature Extraction", "Data Models" ], "authors": [ { "affiliation": "University of Central,Department of Electrical and Computer Engineering,Florida,USA", "fullName": "Umar Khalid", "givenName": "Umar", "surname": "Khalid", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Central,Department of Electrical and Computer Engineering,Florida,USA", "fullName": "Ashkan Esmaeili", "givenName": "Ashkan", "surname": "Esmaeili", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Central,Department of Electrical and Computer Engineering,Florida,USA", "fullName": "Nazmul Karim", "givenName": "Nazmul", "surname": "Karim", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Central,Department of Electrical and Computer Engineering,Florida,USA", "fullName": "Nazanin Rahnavard", "givenName": "Nazanin", "surname": "Rahnavard", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvprw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-06-01T00:00:00", "pubType": "proceedings", "pages": "163-170", "year": "2022", "issn": null, "isbn": "978-1-6654-8739-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "873900a155", "articleId": "1G56N34zTgY", "__typename": "AdjacentArticleType" }, "next": { "fno": "873900a171", "articleId": "1G56U3n8v72", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2021/2812/0/281200d479", "title": "Preservational Learning Improves Self-supervised Medical Image Models by Reconstructing Diverse Contexts", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200d479/1BmHXM56IVy", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200j485", "title": "Solving Inefficiency of Self-supervised Representation Learning", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200j485/1BmIxdNww3m", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200j929", "title": "Contrasting Contrastive Self-Supervised Representation Learning Pipelines", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200j929/1BmKBTXUH4c", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900c836", "title": "Class-wise Thresholding for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900c836/1G56Ht7f7jO", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2022/8563/0/09859802", "title": "Self-Supervised Video Representation Learning with Motion-Contrastive Perception", "doi": null, "abstractUrl": "/proceedings-article/icme/2022/09859802/1G9DTS2aMRq", "parentPublication": { "id": "proceedings/icme/2022/8563/0", "title": "2022 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600j717", "title": "Hierarchical Self-supervised Representation Learning for Movie Understanding", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600j717/1H0NMsTj7kQ", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f500", "title": "No Shifted Augmentations (NSA): compact distributions for robust self-supervised Anomaly Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f500/1L8qwtaPRbq", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bigcomp/2023/7578/0/757800a195", "title": "Attention Masking for Improved Near Out-of-Distribution Image Detection", "doi": null, "abstractUrl": "/proceedings-article/bigcomp/2023/757800a195/1LFLD4mzJ9m", "parentPublication": { "id": "proceedings/bigcomp/2023/7578/0", "title": "2023 IEEE International Conference on Big Data and Smart Computing (BigComp)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ase/2020/6768/0/676800b041", "title": "Cats Are Not Fish: Deep Learning Testing Calls for Out-Of-Distribution Awareness", "doi": null, "abstractUrl": "/proceedings-article/ase/2020/676800b041/1pP3K26HxE4", "parentPublication": { "id": "proceedings/ase/2020/6768/0", "title": "2020 35th IEEE/ACM International Conference on Automated Software Engineering (ASE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2021/0191/0/019100d317", "title": "SOoD: Self-Supervised Out-of-Distribution Detection Under Domain Shift for Multi-Class Colorectal Cancer Tissue Types", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2021/019100d317/1yNiFB25sJi", "parentPublication": { "id": "proceedings/iccvw/2021/0191/0", "title": "2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1H1gVMlkl32", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1H1jN1lNZM4", "doi": "10.1109/CVPR52688.2022.01639", "title": "Weakly Supervised Semantic Segmentation using Out-of-Distribution Data", "normalizedTitle": "Weakly Supervised Semantic Segmentation using Out-of-Distribution Data", "abstract": "Weakly supervised semantic segmentation (WSSS) methods are often built on pixel-level localization maps obtained from a classifier. However, training on class labels only, classifiers suffer from the spurious correlation between fore-ground and background cues (e.g. train and rail), fundamentally bounding the performance of WSSS. There have been previous endeavors to address this issue with additional supervision. We propose a novel source of information to distinguish foreground from the background: Out-of-Distribution (OoD) data, or images devoid of foreground object classes. In particular, we utilize the hard OoDs that the classifier is likely to make false-positive predictions. These samples typically carry key visual features on the background (e.g. rail) that the classifiers often confuse as foreground (e.g. train), so these cues let classifiers correctly suppress spurious background cues. Acquiring such hard OoDs does not require an extensive amount of annotation efforts; it only incurs a few additional image-level labeling costs on top of the original efforts to collect class labels. We propose a method, W-OoD, for utilizing the hard OoDs. W-OoD achieves state-of-the-art performance on Pascal VOC 2012. The code is available at: https://github.com/naver-ai/w-ood.", "abstracts": [ { "abstractType": "Regular", "content": "Weakly supervised semantic segmentation (WSSS) methods are often built on pixel-level localization maps obtained from a classifier. However, training on class labels only, classifiers suffer from the spurious correlation between fore-ground and background cues (e.g. train and rail), fundamentally bounding the performance of WSSS. There have been previous endeavors to address this issue with additional supervision. We propose a novel source of information to distinguish foreground from the background: Out-of-Distribution (OoD) data, or images devoid of foreground object classes. In particular, we utilize the hard OoDs that the classifier is likely to make false-positive predictions. These samples typically carry key visual features on the background (e.g. rail) that the classifiers often confuse as foreground (e.g. train), so these cues let classifiers correctly suppress spurious background cues. Acquiring such hard OoDs does not require an extensive amount of annotation efforts; it only incurs a few additional image-level labeling costs on top of the original efforts to collect class labels. We propose a method, W-OoD, for utilizing the hard OoDs. W-OoD achieves state-of-the-art performance on Pascal VOC 2012. The code is available at: https://github.com/naver-ai/w-ood.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Weakly supervised semantic segmentation (WSSS) methods are often built on pixel-level localization maps obtained from a classifier. However, training on class labels only, classifiers suffer from the spurious correlation between fore-ground and background cues (e.g. train and rail), fundamentally bounding the performance of WSSS. There have been previous endeavors to address this issue with additional supervision. We propose a novel source of information to distinguish foreground from the background: Out-of-Distribution (OoD) data, or images devoid of foreground object classes. In particular, we utilize the hard OoDs that the classifier is likely to make false-positive predictions. These samples typically carry key visual features on the background (e.g. rail) that the classifiers often confuse as foreground (e.g. train), so these cues let classifiers correctly suppress spurious background cues. Acquiring such hard OoDs does not require an extensive amount of annotation efforts; it only incurs a few additional image-level labeling costs on top of the original efforts to collect class labels. We propose a method, W-OoD, for utilizing the hard OoDs. W-OoD achieves state-of-the-art performance on Pascal VOC 2012. The code is available at: https://github.com/naver-ai/w-ood.", "fno": "694600q6876", "keywords": [ "Feature Extraction", "Image Classification", "Image Segmentation", "Object Detection", "Supervised Learning", "Weakly Supervised Semantic Segmentation", "Out Of Distribution Data", "Pixel Level Localization Maps", "Classifier", "Class Labels", "Spurious Correlation", "Foreground Object Classes", "Spurious Background Cues", "W Oo D Achieves State Of The Art Performance", "Image Level Labeling Costs", "Visual Features", "Rails", "Training", "Location Awareness", "Visualization", "Image Segmentation", "Image Analysis", "Shape", "Scene Analysis And Understanding Segmentation", "Grouping And Shape Analysis" ], "authors": [ { "affiliation": "Seoul National University,Department of Electrical and Computer Engineering", "fullName": "Jungbeom Lee", "givenName": "Jungbeom", "surname": "Lee", "__typename": "ArticleAuthorType" }, { "affiliation": "NAVER AI Lab", "fullName": "Seong Joon Oh", "givenName": "Seong Joon", "surname": "Oh", "__typename": "ArticleAuthorType" }, { "affiliation": "NAVER AI Lab", "fullName": "Sangdoo Yun", "givenName": "Sangdoo", "surname": "Yun", "__typename": "ArticleAuthorType" }, { "affiliation": "Sogang University,Department of Computer Science and Engineering", "fullName": "Junsuk Choe", "givenName": "Junsuk", "surname": "Choe", "__typename": "ArticleAuthorType" }, { "affiliation": "Seoul National University,Department of Electrical and Computer Engineering", "fullName": "Eunji Kim", "givenName": "Eunji", "surname": "Kim", "__typename": "ArticleAuthorType" }, { "affiliation": "Seoul National University,Department of Electrical and Computer Engineering", "fullName": "Sungroh Yoon", "givenName": "Sungroh", "surname": "Yoon", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-06-01T00:00:00", "pubType": "proceedings", "pages": "16876-16885", "year": "2022", "issn": null, "isbn": "978-1-6654-6946-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1H1jMY4QNLW", "name": "pcvpr202269460-09879885s1-mm_694600q6876.zip", "size": "2.06 MB", "location": "https://www.computer.org/csdl/api/v1/extra/pcvpr202269460-09879885s1-mm_694600q6876.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "694600q6865", "articleId": "1H1i5LOMmru", "__typename": "AdjacentArticleType" }, "next": { "fno": "694600q6886", "articleId": "1H1mpqcpVWo", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/wacv/2022/0915/0/091500c653", "title": "Inferring the Class Conditional Response Map for Weakly Supervised Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500c653/1B13f1ogUkE", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/5555/01/09760057", "title": "Looking Beyond Single Images for Weakly Supervised Semantic Segmentation Learning", "doi": null, "abstractUrl": "/journal/tp/5555/01/09760057/1CHszUuMU00", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/6.946E273", "title": "Beyond Semantic to Instance Segmentation: Weakly-Supervised Instance Segmentation via Semantic Knowledge Transfer and Self-Refinement", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/6.946E273/1H0KZg9VpaE", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600a979", "title": "C<sup>2</sup> AM: Contrastive learning of Class-agnostic Activation Map for Weakly Supervised Object Localization and Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600a979/1H0N20PbP1u", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/6.946E283", "title": "Self-supervised Image-specific Prototype Exploration for Weakly Supervised Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/6.946E283/1H1jSqRFK24", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icftic/2022/2195/0/10075115", "title": "Attention Based Object Localization for Weakly Supervised Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/icftic/2022/10075115/1LRlj96lkR2", "parentPublication": { "id": "proceedings/icftic/2022/2195/0", "title": "2022 4th International Conference on Frontiers Technology of Information and Computer (ICFTIC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/7.168E287", "title": "Learning Integral Objects With Intra-Class Discriminator for Weakly-Supervised Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/7.168E287/1m3nRMNdemc", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2021/3864/0/09428190", "title": "Weakly-Supervised Attribute Segmentation", "doi": null, "abstractUrl": "/proceedings-article/icme/2021/09428190/1uim4ZS0zba", "parentPublication": { "id": "proceedings/icme/2021/3864/0", "title": "2021 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900f491", "title": "Railroad is not a Train: Saliency as Pseudo-pixel Supervision for Weakly Supervised Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900f491/1yeJx33VrUY", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900g909", "title": "Background-Aware Pooling and Noise-Aware Loss for Weakly-Supervised Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900g909/1yeLcbdNrOg", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1KxUhhFgzlK", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "acronym": "wacv", "groupId": "1000040", "volume": "0", "displayVolume": "0", "year": "2023", "__typename": "ProceedingType" }, "article": { "id": "1L8qqu8Q3OU", "doi": "10.1109/WACV56688.2023.00263", "title": "Heatmap-based Out-of-Distribution Detection", "normalizedTitle": "Heatmap-based Out-of-Distribution Detection", "abstract": "Our work investigates out-of-distribution (OOD) detection as a neural network output explanation problem. We learn a heatmap representation for detecting OOD images while visualizing in- and out-of-distribution image regions at the same time. Given a trained and fixed classifier, we train a decoder neural network to produce heatmaps with zero response for in-distribution samples and high response heatmaps for OOD samples, based on the classifier features and the class prediction. Our main innovation lies in the heatmap definition for an OOD sample, as the normalized difference from the closest in-distribution sample. The heatmap serves as a margin to distinguish between in- and out-of-distribution samples. Our approach generates the heatmaps not only for OOD detection, but also to indicates in- and out-of-distribution regions of the input image. In our evaluations, our approach mostly outperforms the prior work on fixed classifiers, trained on CIFAR-10, CIFAR-100 and Tiny ImageNet. The code is publicly available at: https://github.com/jhornauer/heatmap_ood.", "abstracts": [ { "abstractType": "Regular", "content": "Our work investigates out-of-distribution (OOD) detection as a neural network output explanation problem. We learn a heatmap representation for detecting OOD images while visualizing in- and out-of-distribution image regions at the same time. Given a trained and fixed classifier, we train a decoder neural network to produce heatmaps with zero response for in-distribution samples and high response heatmaps for OOD samples, based on the classifier features and the class prediction. Our main innovation lies in the heatmap definition for an OOD sample, as the normalized difference from the closest in-distribution sample. The heatmap serves as a margin to distinguish between in- and out-of-distribution samples. Our approach generates the heatmaps not only for OOD detection, but also to indicates in- and out-of-distribution regions of the input image. In our evaluations, our approach mostly outperforms the prior work on fixed classifiers, trained on CIFAR-10, CIFAR-100 and Tiny ImageNet. The code is publicly available at: https://github.com/jhornauer/heatmap_ood.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Our work investigates out-of-distribution (OOD) detection as a neural network output explanation problem. We learn a heatmap representation for detecting OOD images while visualizing in- and out-of-distribution image regions at the same time. Given a trained and fixed classifier, we train a decoder neural network to produce heatmaps with zero response for in-distribution samples and high response heatmaps for OOD samples, based on the classifier features and the class prediction. Our main innovation lies in the heatmap definition for an OOD sample, as the normalized difference from the closest in-distribution sample. The heatmap serves as a margin to distinguish between in- and out-of-distribution samples. Our approach generates the heatmaps not only for OOD detection, but also to indicates in- and out-of-distribution regions of the input image. In our evaluations, our approach mostly outperforms the prior work on fixed classifiers, trained on CIFAR-10, CIFAR-100 and Tiny ImageNet. The code is publicly available at: https://github.com/jhornauer/heatmap_ood.", "fno": "934600c602", "keywords": [ "Deep Learning Artificial Intelligence", "Explanation", "Image Classification", "Learning Artificial Intelligence", "Neural Nets", "Pattern Classification", "Classifier Features", "Decoder Neural Network", "Detecting OOD Images", "Fixed Classifiers", "Heatmap Definition", "Heatmap Representation", "Heatmap Based Out Of Distribution Detection", "High Response Heatmaps", "In Distribution Sample", "Input Image", "Neural Network Output Explanation Problem", "OOD Detection", "OOD Sample", "Out Of Distribution Image Regions", "Out Of Distribution Regions", "Out Of Distribution Samples", "Trained Fixed Classifier", "Zero Response", "Heating Systems", "Visualization", "Technological Innovation", "Computer Vision", "Codes", "Neural Networks", "Decoding", "Algorithms Explainable", "Fair", "Accountable", "Privacy Preserving", "Ethical Computer Vision" ], "authors": [ { "affiliation": "Ulm University,Germany", "fullName": "Julia Hornauer", "givenName": "Julia", "surname": "Hornauer", "__typename": "ArticleAuthorType" }, { "affiliation": "Ulm University,Germany", "fullName": "Vasileios Belagiannis", "givenName": "Vasileios", "surname": "Belagiannis", "__typename": "ArticleAuthorType" } ], "idPrefix": "wacv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2023-01-01T00:00:00", "pubType": "proceedings", "pages": "2602-2611", "year": "2023", "issn": null, "isbn": "978-1-6654-9346-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "934600c591", "articleId": "1KxVasKgAmY", "__typename": "AdjacentArticleType" }, "next": { "fno": "934600c612", "articleId": "1L6LAu5ndXG", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2021/2812/0/281200i281", "title": "Semantically Coherent Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200i281/1BmFq1tAb16", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900d849", "title": "Out-Of-Distribution Detection In Unsupervised Continual Learning", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900d849/1G4F61c26GI", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900a163", "title": "RODD: A Self-Supervised Approach for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900a163/1G56R8MQAwg", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600e723", "title": "Deep Hybrid Models for Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600e723/1H1jrFQvDNu", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f540", "title": "Out-of-Distribution Detection with Reconstruction Error and Typicality-based Penalty", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f540/1KxVfdAFybS", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bigcomp/2023/7578/0/757800a195", "title": "Attention Masking for Improved Near Out-of-Distribution Image Detection", "doi": null, "abstractUrl": "/proceedings-article/bigcomp/2023/757800a195/1LFLD4mzJ9m", "parentPublication": { "id": "proceedings/bigcomp/2023/7578/0", "title": "2023 IEEE International Conference on Big Data and Smart Computing (BigComp)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800h091", "title": "Distribution-Aware Coordinate Representation for Human Pose Estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800h091/1m3odsWY65O", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/11/09511827", "title": "Heatmap Regression via Randomized Rounding", "doi": null, "abstractUrl": "/journal/tp/2022/11/09511827/1vYRGdzUT5u", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/mipr/2021/1865/0/186500a377", "title": "Respective Volumetric Heatmap Autoencoder for Multi-Person 3D Pose Estimation", "doi": null, "abstractUrl": "/proceedings-article/mipr/2021/186500a377/1xPsokwn38Y", "parentPublication": { "id": "proceedings/mipr/2021/1865/0", "title": "2021 IEEE 4th International Conference on Multimedia Information Processing and Retrieval (MIPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900i706", "title": "MOS: Towards Scaling Out-of-distribution Detection for Large Semantic Space", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900i706/1yeMjKnNiEw", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1KxUhhFgzlK", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "acronym": "wacv", "groupId": "1000040", "volume": "0", "displayVolume": "0", "year": "2023", "__typename": "ProceedingType" }, "article": { "id": "1L8qr7oAN44", "doi": "10.1109/WACV56688.2023.00267", "title": "Hyperdimensional Feature Fusion for Out-of-Distribution Detection", "normalizedTitle": "Hyperdimensional Feature Fusion for Out-of-Distribution Detection", "abstract": "We introduce powerful ideas from Hyperdimensional Computing into the challenging field of Out-of-Distribution (OOD) detection. In contrast to most existing works that perform OOD detection based on only a single layer of a neural network, we use similarity-preserving semi-orthogonal projection matrices to project the feature maps from multiple layers into a common vector space. By repeatedly applying the bundling operation &#x2295;, we create expressive class-specific descriptor vectors for all in-distribution classes. At test time, a simple and efficient cosine similarity calculation between descriptor vectors consistently identifies OOD samples with competitive performance to the current state-of-the-art whilst being significantly faster. We show that our method is orthogonal to recent state-of-the-art OOD detectors and can be combined with them to further improve upon the performance.", "abstracts": [ { "abstractType": "Regular", "content": "We introduce powerful ideas from Hyperdimensional Computing into the challenging field of Out-of-Distribution (OOD) detection. In contrast to most existing works that perform OOD detection based on only a single layer of a neural network, we use similarity-preserving semi-orthogonal projection matrices to project the feature maps from multiple layers into a common vector space. By repeatedly applying the bundling operation &#x2295;, we create expressive class-specific descriptor vectors for all in-distribution classes. At test time, a simple and efficient cosine similarity calculation between descriptor vectors consistently identifies OOD samples with competitive performance to the current state-of-the-art whilst being significantly faster. We show that our method is orthogonal to recent state-of-the-art OOD detectors and can be combined with them to further improve upon the performance.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We introduce powerful ideas from Hyperdimensional Computing into the challenging field of Out-of-Distribution (OOD) detection. In contrast to most existing works that perform OOD detection based on only a single layer of a neural network, we use similarity-preserving semi-orthogonal projection matrices to project the feature maps from multiple layers into a common vector space. By repeatedly applying the bundling operation ⊕, we create expressive class-specific descriptor vectors for all in-distribution classes. At test time, a simple and efficient cosine similarity calculation between descriptor vectors consistently identifies OOD samples with competitive performance to the current state-of-the-art whilst being significantly faster. We show that our method is orthogonal to recent state-of-the-art OOD detectors and can be combined with them to further improve upon the performance.", "fno": "934600c643", "keywords": [ "Feature Extraction", "Learning Artificial Intelligence", "Matrix Algebra", "Vectors", "Bundling Operation", "Challenging Field", "Common Vector Space", "Competitive Performance", "Current State Of The Art Whilst", "Efficient Cosine Similarity Calculation", "Expressive Class Specific Descriptor Vectors", "Feature Maps", "Hyperdimensional Computing", "Hyperdimensional Feature Fusion", "In Distribution Classes", "Multiple Layers", "Neural Network", "OOD Detection", "Out Of Distribution Detection", "Powerful Ideas", "Recent State Of The Art OOD Detectors", "Similarity Preserving Semiorthogonal Projection Matrices", "Simple Cosine Similarity Calculation", "Single Layer", "Visualization", "Computer Vision", "Sensitivity", "Neural Networks", "Detectors", "Feature Extraction", "Computational Efficiency", "Algorithms Image Recognition And Understanding Object Detection", "Categorization", "Segmentation" ], "authors": [ { "affiliation": "Queensland University of Technology,Brisbane,QLD,Australia,4000", "fullName": "Samuel Wilson", "givenName": "Samuel", "surname": "Wilson", "__typename": "ArticleAuthorType" }, { "affiliation": "Queensland University of Technology,Brisbane,QLD,Australia,4000", "fullName": "Tobias Fischer", "givenName": "Tobias", "surname": "Fischer", "__typename": "ArticleAuthorType" }, { "affiliation": "Queensland University of Technology,Brisbane,QLD,Australia,4000", "fullName": "Niko Sünderhauf", "givenName": "Niko", "surname": "Sünderhauf", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Adelaide,Adelaide,SA,Australia,5005", "fullName": "Feras Dayoub", "givenName": "Feras", "surname": "Dayoub", "__typename": "ArticleAuthorType" } ], "idPrefix": "wacv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2023-01-01T00:00:00", "pubType": "proceedings", "pages": "2643-2653", "year": "2023", "issn": null, "isbn": "978-1-6654-9346-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "934600c633", "articleId": "1KxVwjP0Lks", "__typename": "AdjacentArticleType" }, "next": { "fno": "934600c654", "articleId": "1KxV0YRiBYA", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2021/2812/0/281200p5681", "title": "Triggering Failures: Out-Of-Distribution detection by learning from local adversarial attacks in Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200p5681/1BmFVHaA1l6", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200b133", "title": "CODEs: Chamfer Out-of-Distribution Examples against Overconfidence Issue", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200b133/1BmFaDYHtgA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900d849", "title": "Out-Of-Distribution Detection In Unsupervised Continual Learning", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900d849/1G4F61c26GI", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900c836", "title": "Class-wise Thresholding for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900c836/1G56Ht7f7jO", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600e911", "title": "ViM: Out-Of-Distribution with Virtual-logit Matching", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600e911/1H0LoDIYyB2", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2022/9062/0/09956184", "title": "Episodic Projection Network for Out-of-Distribution Detection in Few-shot Learning", "doi": null, "abstractUrl": "/proceedings-article/icpr/2022/09956184/1IHqfHmafKM", "parentPublication": { "id": "proceedings/icpr/2022/9062/0", "title": "2022 26th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f520", "title": "Mixture Outlier Exposure: Towards Out-of-Distribution Detection in Fine-grained Environments", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f520/1KxUKOwyJJS", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600c602", "title": "Heatmap-based Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600c602/1L8qqu8Q3OU", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900p5308", "title": "MOOD: Multi-level Out-of-distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900p5308/1yeHIgRMMgg", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2021/2354/0/235400a409", "title": "GCOOD: A Generic Coupled Out-of-Distribution Detector for Robust Classification", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2021/235400a409/1zurshXlLR6", "parentPublication": { "id": "proceedings/sibgrapi/2021/2354/0", "title": "2021 34th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1hQqfuoOyHu", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2019", "__typename": "ProceedingType" }, "article": { "id": "1hVl9mSRtfO", "doi": "10.1109/ICCV.2019.00961", "title": "Unsupervised Out-of-Distribution Detection by Maximum Classifier Discrepancy", "normalizedTitle": "Unsupervised Out-of-Distribution Detection by Maximum Classifier Discrepancy", "abstract": "Since deep learning models have been implemented in many commercial applications, it is important to detect out-of-distribution (OOD) inputs correctly to maintain the performance of the models, ensure the quality of the collected data, and prevent the applications from being used for other-than-intended purposes. In this work, we propose a two-head deep convolutional neural network (CNN) and maximize the discrepancy between the two classifiers to detect OOD inputs. We train a two-head CNN consisting of one common feature extractor and two classifiers which have different decision boundaries but can classify in-distribution (ID) samples correctly. Unlike previous methods, we also utilize unlabeled data for unsupervised training and we use these unlabeled data to maximize the discrepancy between the decision boundaries of two classifiers to push OOD samples outside the manifold of the in-distribution (ID) samples, which enables us to detect OOD samples that are far from the support of the ID samples. Overall, our approach significantly outperforms other state-of-the-art methods on several OOD detection benchmarks and two cases of real-world simulation.", "abstracts": [ { "abstractType": "Regular", "content": "Since deep learning models have been implemented in many commercial applications, it is important to detect out-of-distribution (OOD) inputs correctly to maintain the performance of the models, ensure the quality of the collected data, and prevent the applications from being used for other-than-intended purposes. In this work, we propose a two-head deep convolutional neural network (CNN) and maximize the discrepancy between the two classifiers to detect OOD inputs. We train a two-head CNN consisting of one common feature extractor and two classifiers which have different decision boundaries but can classify in-distribution (ID) samples correctly. Unlike previous methods, we also utilize unlabeled data for unsupervised training and we use these unlabeled data to maximize the discrepancy between the decision boundaries of two classifiers to push OOD samples outside the manifold of the in-distribution (ID) samples, which enables us to detect OOD samples that are far from the support of the ID samples. Overall, our approach significantly outperforms other state-of-the-art methods on several OOD detection benchmarks and two cases of real-world simulation.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Since deep learning models have been implemented in many commercial applications, it is important to detect out-of-distribution (OOD) inputs correctly to maintain the performance of the models, ensure the quality of the collected data, and prevent the applications from being used for other-than-intended purposes. In this work, we propose a two-head deep convolutional neural network (CNN) and maximize the discrepancy between the two classifiers to detect OOD inputs. We train a two-head CNN consisting of one common feature extractor and two classifiers which have different decision boundaries but can classify in-distribution (ID) samples correctly. Unlike previous methods, we also utilize unlabeled data for unsupervised training and we use these unlabeled data to maximize the discrepancy between the decision boundaries of two classifiers to push OOD samples outside the manifold of the in-distribution (ID) samples, which enables us to detect OOD samples that are far from the support of the ID samples. Overall, our approach significantly outperforms other state-of-the-art methods on several OOD detection benchmarks and two cases of real-world simulation.", "fno": "480300j517", "keywords": [ "Convolutional Neural Nets", "Feature Extraction", "Learning Artificial Intelligence", "Pattern Classification", "Out Of Distribution Detection", "Maximum Classifier Discrepancy", "Deep Learning Models", "Commercial Applications", "Deep Convolutional Neural Network", "CNN", "Feature Extractor", "Decision Boundaries", "Unlabeled Data", "Unsupervised Training", "ID Samples", "OOD Detection Benchmarks", "Training", "Feature Extraction", "Neural Networks", "Data Models", "Entropy", "Manifolds", "Task Analysis" ], "authors": [ { "affiliation": "The University of Tokyo", "fullName": "Qing Yu", "givenName": "Qing", "surname": "Yu", "__typename": "ArticleAuthorType" }, { "affiliation": "The University of Tokyo", "fullName": "Kiyoharu Aizawa", "givenName": "Kiyoharu", "surname": "Aizawa", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2019-10-01T00:00:00", "pubType": "proceedings", "pages": "9517-9525", "year": "2019", "issn": null, "isbn": "978-1-7281-4803-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "480300j507", "articleId": "1hQqpTLgM1O", "__typename": "AdjacentArticleType" }, "next": { "fno": "480300j526", "articleId": "1hQqk6ejlpm", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2018/6420/0/642000d723", "title": "Maximum Classifier Discrepancy for Unsupervised Domain Adaptation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000d723/17D45VsBTVR", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200i281", "title": "Semantically Coherent Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200i281/1BmFq1tAb16", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2023/03/09772946", "title": "Maximum Structural Generation Discrepancy for Unsupervised Domain Adaptation", "doi": null, "abstractUrl": "/journal/tp/2023/03/09772946/1DhYymTKSGY", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600e911", "title": "ViM: Out-Of-Distribution with Virtual-logit Matching", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600e911/1H0LoDIYyB2", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600t9195", "title": "Neural Mean Discrepancy for Efficient Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600t9195/1H1mfixU9JC", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/5555/01/09987694", "title": "Revealing the Distributional Vulnerability of Discriminators by Implicit Generators", "doi": null, "abstractUrl": "/journal/tp/5555/01/09987694/1J7RLU1AUAU", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600c602", "title": "Heatmap-based Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600c602/1L8qqu8Q3OU", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2021/3864/0/09428418", "title": "Unsupervised Domain Adaptation VIA Cluster Alignment with Maximum Classifier Discrepancy", "doi": null, "abstractUrl": "/proceedings-article/icme/2021/09428418/1uim2ilgDyo", "parentPublication": { "id": "proceedings/icme/2021/3864/0", "title": "2021 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900d936", "title": "Cross-Domain Gradient Discrepancy Minimization for Unsupervised Domain Adaptation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900d936/1yeJQKEWvBu", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900j447", "title": "Out-of-Distribution Detection Using Union of 1-Dimensional Subspaces", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900j447/1yeKwz9Z2Uw", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1tmhi3ly74c", "title": "2020 25th International Conference on Pattern Recognition (ICPR)", "acronym": "icpr", "groupId": "1000545", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1tmiAPeMOOs", "doi": "10.1109/ICPR48806.2021.9412612", "title": "Boundary Optimised Samples Training for Detecting Out-of-Distribution Images", "normalizedTitle": "Boundary Optimised Samples Training for Detecting Out-of-Distribution Images", "abstract": "This paper presents a new approach to the problem of detecting out-of-distribution (OOD) inputs in image classifications with deep convolutional networks. We leverage so-called boundary samples to enforce low confidence (maximum softmax probabilities) for inputs far away from the training data. In particular, we propose the boundary optimised samples (named BoS) training algorithm for generating them. Unlike existing approaches, it does not require extra generative adversarial network, but achieves the goal by simply back propagating the gradient of an appropriately designed loss function to the input samples. At the end of the BoS training, all the boundary samples are in principle located on a specific level hypersurface with respect to the designed loss. Our contributions are i) the BoS training as an efficient alternative to generate boundary samples, ii) a robust algorithm therewith to enforce low confidence for OOD samples, and iii) experiments demonstrating improved OOD detection over the baseline. We show the performance using standard datasets for training and different test sets including Fashion MNIST, EMNIST, SVHN, and CIFAR-100, preceded by evaluations with a synthetic 2-dimensional dataset that provide an insight for the new procedure.", "abstracts": [ { "abstractType": "Regular", "content": "This paper presents a new approach to the problem of detecting out-of-distribution (OOD) inputs in image classifications with deep convolutional networks. We leverage so-called boundary samples to enforce low confidence (maximum softmax probabilities) for inputs far away from the training data. In particular, we propose the boundary optimised samples (named BoS) training algorithm for generating them. Unlike existing approaches, it does not require extra generative adversarial network, but achieves the goal by simply back propagating the gradient of an appropriately designed loss function to the input samples. At the end of the BoS training, all the boundary samples are in principle located on a specific level hypersurface with respect to the designed loss. Our contributions are i) the BoS training as an efficient alternative to generate boundary samples, ii) a robust algorithm therewith to enforce low confidence for OOD samples, and iii) experiments demonstrating improved OOD detection over the baseline. We show the performance using standard datasets for training and different test sets including Fashion MNIST, EMNIST, SVHN, and CIFAR-100, preceded by evaluations with a synthetic 2-dimensional dataset that provide an insight for the new procedure.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper presents a new approach to the problem of detecting out-of-distribution (OOD) inputs in image classifications with deep convolutional networks. We leverage so-called boundary samples to enforce low confidence (maximum softmax probabilities) for inputs far away from the training data. In particular, we propose the boundary optimised samples (named BoS) training algorithm for generating them. Unlike existing approaches, it does not require extra generative adversarial network, but achieves the goal by simply back propagating the gradient of an appropriately designed loss function to the input samples. At the end of the BoS training, all the boundary samples are in principle located on a specific level hypersurface with respect to the designed loss. Our contributions are i) the BoS training as an efficient alternative to generate boundary samples, ii) a robust algorithm therewith to enforce low confidence for OOD samples, and iii) experiments demonstrating improved OOD detection over the baseline. We show the performance using standard datasets for training and different test sets including Fashion MNIST, EMNIST, SVHN, and CIFAR-100, preceded by evaluations with a synthetic 2-dimensional dataset that provide an insight for the new procedure.", "fno": "09412612", "keywords": [ "Backpropagation", "Convolutional Neural Nets", "Deep Learning Artificial Intelligence", "Gradient Methods", "Image Classification", "Optimisation", "Probability", "Maximum Softmax Probabilities", "Bo S Training", "OOD Detection", "Boundary Optimised Samples Training", "Image Classifications", "Deep Convolutional Networks", "Out Of Distribution Image Detection", "Fashion MNIST Dataset", "EMNIST Dataset", "SVHN Dataset", "CIFAR 100 Dataset", "Backpropagation", "Training", "Measurement", "Toy Manufacturing Industry", "Training Data", "Data Visualization", "Benchmark Testing", "Propagation Losses" ], "authors": [ { "affiliation": "Nordetect ApS", "fullName": "Luca Marson", "givenName": "Luca", "surname": "Marson", "__typename": "ArticleAuthorType" }, { "affiliation": "KTH Royal Institute of Technology,Division of Robotics, Perception, and Learning,Stockholm,Sweden", "fullName": "Vladimir Li", "givenName": "Vladimir", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "KTH Royal Institute of Technology,Division of Robotics, Perception, and Learning,Stockholm,Sweden", "fullName": "Atsuto Maki", "givenName": "Atsuto", "surname": "Maki", "__typename": "ArticleAuthorType" } ], "idPrefix": "icpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-01-01T00:00:00", "pubType": "proceedings", "pages": "10486-10492", "year": "2021", "issn": "1051-4651", "isbn": "978-1-7281-8808-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "09413161", "articleId": "1tmhNHyik7e", "__typename": "AdjacentArticleType" }, "next": { "fno": "09412732", "articleId": "1tmjQq0aT0Q", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icisce/2016/2535/0/2535a194", "title": "An Imbalanced Data Classification Method Driven by Boundary Samples-Boundary-Boost", "doi": null, "abstractUrl": "/proceedings-article/icisce/2016/2535a194/12OmNx8fimK", "parentPublication": { "id": "proceedings/icisce/2016/2535/0", "title": "2016 3rd International Conference on Information Science and Control Engineering (ICISCE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200b133", "title": "CODEs: Chamfer Out-of-Distribution Examples against Overconfidence Issue", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200b133/1BmFaDYHtgA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900a163", "title": "RODD: A Self-Supervised Approach for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900a163/1G56R8MQAwg", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600t9195", "title": "Neural Mean Discrepancy for Efficient Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600t9195/1H1mfixU9JC", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600e709", "title": "Augmentation by Counterfactual Explanation -Fixing an Overconfident Classifier", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600e709/1KxUPDDefEQ", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300j517", "title": "Unsupervised Out-of-Distribution Detection by Maximum Classifier Discrepancy", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300j517/1hVl9mSRtfO", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08994105", "title": "OoDAnalyzer: Interactive Analysis of Out-of-Distribution Samples", "doi": null, "abstractUrl": "/journal/tg/2021/07/08994105/1hkQR7Os8SI", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800k0948", "title": "Generalized ODIN: Detecting Out-of-Distribution Image Without Learning From Out-of-Distribution Data", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800k0948/1m3ofcCTYha", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/apsec/2020/9553/0/955300a266", "title": "An Empirical Study on Robustness of DNNs with Out-of-Distribution Awareness", "doi": null, "abstractUrl": "/proceedings-article/apsec/2020/955300a266/1rCgF3Lc6mA", "parentPublication": { "id": "proceedings/apsec/2020/9553/0", "title": "2020 27th Asia-Pacific Software Engineering Conference (APSEC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900j447", "title": "Out-of-Distribution Detection Using Union of 1-Dimensional Subspaces", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900j447/1yeKwz9Z2Uw", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1wzs0vrjyWQ", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "acronym": "cvprw", "groupId": "1001809", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1yXsVNgnUA0", "doi": "10.1109/CVPRW53098.2021.00367", "title": "Sample-free white-box out-of-distribution detection for deep learning", "normalizedTitle": "Sample-free white-box out-of-distribution detection for deep learning", "abstract": "Being able to detect irrelevant test examples with respect to deployed deep learning models is paramount to properly and safely using them. In this paper, we address the problem of rejecting such out-of-distribution (OOD) samples in a fully sample-free way, i.e., without requiring any access to in- distribution or OOD samples. We propose several indicators which can be computed alongside the prediction with little additional cost, assuming white-box access to the network. These indicators prove useful, stable and complementary for OOD detection on frequently-used architectures. We also introduce a surprisingly simple, yet effective summary OOD indicator. This indicator is shown to perform well across several networks and datasets and can furthermore be easily tuned as soon as samples become available. Lastly, we discuss how to exploit this summary in real-world settings.", "abstracts": [ { "abstractType": "Regular", "content": "Being able to detect irrelevant test examples with respect to deployed deep learning models is paramount to properly and safely using them. In this paper, we address the problem of rejecting such out-of-distribution (OOD) samples in a fully sample-free way, i.e., without requiring any access to in- distribution or OOD samples. We propose several indicators which can be computed alongside the prediction with little additional cost, assuming white-box access to the network. These indicators prove useful, stable and complementary for OOD detection on frequently-used architectures. We also introduce a surprisingly simple, yet effective summary OOD indicator. This indicator is shown to perform well across several networks and datasets and can furthermore be easily tuned as soon as samples become available. Lastly, we discuss how to exploit this summary in real-world settings.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Being able to detect irrelevant test examples with respect to deployed deep learning models is paramount to properly and safely using them. In this paper, we address the problem of rejecting such out-of-distribution (OOD) samples in a fully sample-free way, i.e., without requiring any access to in- distribution or OOD samples. We propose several indicators which can be computed alongside the prediction with little additional cost, assuming white-box access to the network. These indicators prove useful, stable and complementary for OOD detection on frequently-used architectures. We also introduce a surprisingly simple, yet effective summary OOD indicator. This indicator is shown to perform well across several networks and datasets and can furthermore be easily tuned as soon as samples become available. Lastly, we discuss how to exploit this summary in real-world settings.", "fno": "489900d285", "keywords": [ "Deep Learning", "Computer Vision", "Filtering", "Computational Modeling", "Conferences", "Computer Architecture", "Data Models" ], "authors": [ { "affiliation": "University of Liege", "fullName": "Jean-Michel Begon", "givenName": "Jean-Michel", "surname": "Begon", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Liege", "fullName": "Pierre Geurts", "givenName": "Pierre", "surname": "Geurts", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvprw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-06-01T00:00:00", "pubType": "proceedings", "pages": "3285-3294", "year": "2021", "issn": null, "isbn": "978-1-6654-4899-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1yZ54Ugm98s", "name": "pcvprw202148990-09522831s1-mm_489900d285.zip", "size": "2.51 MB", "location": "https://www.computer.org/csdl/api/v1/extra/pcvprw202148990-09522831s1-mm_489900d285.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "489900d275", "articleId": "1yVzXhfh7dS", "__typename": "AdjacentArticleType" }, "next": { "fno": "489900d295", "articleId": "1yVA4C5LsPu", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2021/2812/0/281200b133", "title": "CODEs: Chamfer Out-of-Distribution Examples against Overconfidence Issue", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200b133/1BmFaDYHtgA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200f108", "title": "Entropy Maximization and Meta Classification for Out-of-Distribution Detection in Semantic Segmentation", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200f108/1BmKpUcBV5e", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900d849", "title": "Out-Of-Distribution Detection In Unsupervised Continual Learning", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900d849/1G4F61c26GI", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900c836", "title": "Class-wise Thresholding for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900c836/1G56Ht7f7jO", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600e911", "title": "ViM: Out-Of-Distribution with Virtual-logit Matching", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600e911/1H0LoDIYyB2", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f520", "title": "Mixture Outlier Exposure: Towards Out-of-Distribution Detection in Fine-grained Environments", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f520/1KxUKOwyJJS", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600c602", "title": "Heatmap-based Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600c602/1L8qqu8Q3OU", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600c643", "title": "Hyperdimensional Feature Fusion for Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600c643/1L8qr7oAN44", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bigcomp/2023/7578/0/757800a195", "title": "Attention Masking for Improved Near Out-of-Distribution Image Detection", "doi": null, "abstractUrl": "/proceedings-article/bigcomp/2023/757800a195/1LFLD4mzJ9m", "parentPublication": { "id": "proceedings/bigcomp/2023/7578/0", "title": "2023 IEEE International Conference on Big Data and Smart Computing (BigComp)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900j447", "title": "Out-of-Distribution Detection Using Union of 1-Dimensional Subspaces", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900j447/1yeKwz9Z2Uw", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1yeHGyRsuys", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1yeKwz9Z2Uw", "doi": "10.1109/CVPR46437.2021.00933", "title": "Out-of-Distribution Detection Using Union of 1-Dimensional Subspaces", "normalizedTitle": "Out-of-Distribution Detection Using Union of 1-Dimensional Subspaces", "abstract": "The goal of out-of-distribution (OOD) detection is to handle the situations where the test samples are drawn from a different distribution than the training data. In this paper, we argue that OOD samples can be detected more easily if the training data is embedded into a low-dimensional space, such that the embedded training samples lie on a union of 1-dimensional subspaces. We show that such embedding of the in-distribution (ID) samples provides us with two main advantages. First, due to compact representation in the feature space, OOD samples are less likely to occupy the same region as the known classes. Second, the first singular vector of ID samples belonging to a 1-dimensional subspace can be used as their robust representative. Motivated by these observations, we train a deep neural network such that the ID samples are embedded onto a union of 1-dimensional subspaces. At the test time, employing sampling techniques used for approximate Bayesian inference in deep learning, input samples are detected as OOD if they occupy the region corresponding to the ID samples with probability 0. Spectral components of the ID samples are used as robust representative of this region. Our method does not have any hyperparameter to be tuned using extra information and it can be applied on different modalities with minimal change. The effectiveness of the proposed method is demonstrated on different benchmark datasets, both in the image and video classification domains.", "abstracts": [ { "abstractType": "Regular", "content": "The goal of out-of-distribution (OOD) detection is to handle the situations where the test samples are drawn from a different distribution than the training data. In this paper, we argue that OOD samples can be detected more easily if the training data is embedded into a low-dimensional space, such that the embedded training samples lie on a union of 1-dimensional subspaces. We show that such embedding of the in-distribution (ID) samples provides us with two main advantages. First, due to compact representation in the feature space, OOD samples are less likely to occupy the same region as the known classes. Second, the first singular vector of ID samples belonging to a 1-dimensional subspace can be used as their robust representative. Motivated by these observations, we train a deep neural network such that the ID samples are embedded onto a union of 1-dimensional subspaces. At the test time, employing sampling techniques used for approximate Bayesian inference in deep learning, input samples are detected as OOD if they occupy the region corresponding to the ID samples with probability 0. Spectral components of the ID samples are used as robust representative of this region. Our method does not have any hyperparameter to be tuned using extra information and it can be applied on different modalities with minimal change. The effectiveness of the proposed method is demonstrated on different benchmark datasets, both in the image and video classification domains.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The goal of out-of-distribution (OOD) detection is to handle the situations where the test samples are drawn from a different distribution than the training data. In this paper, we argue that OOD samples can be detected more easily if the training data is embedded into a low-dimensional space, such that the embedded training samples lie on a union of 1-dimensional subspaces. We show that such embedding of the in-distribution (ID) samples provides us with two main advantages. First, due to compact representation in the feature space, OOD samples are less likely to occupy the same region as the known classes. Second, the first singular vector of ID samples belonging to a 1-dimensional subspace can be used as their robust representative. Motivated by these observations, we train a deep neural network such that the ID samples are embedded onto a union of 1-dimensional subspaces. At the test time, employing sampling techniques used for approximate Bayesian inference in deep learning, input samples are detected as OOD if they occupy the region corresponding to the ID samples with probability 0. Spectral components of the ID samples are used as robust representative of this region. Our method does not have any hyperparameter to be tuned using extra information and it can be applied on different modalities with minimal change. The effectiveness of the proposed method is demonstrated on different benchmark datasets, both in the image and video classification domains.", "fno": "450900j447", "keywords": [ "Bayes Methods", "Deep Learning Artificial Intelligence", "Feature Extraction", "Image Classification", "Image Sampling", "Inference Mechanisms", "Statistical Distributions", "1 Dimensional Subspace", "Sampling Techniques", "Out Of Distribution Detection", "OOD Samples", "Embedded Training Samples", "In Distribution Samples", "Feature Space", "Deep Neural Network Training", "Approximate Bayesian Inference", "Deep Learning", "Probability", "Deep Learning", "Training", "Measurement", "Computer Vision", "Training Data", "Benchmark Testing", "Feature Extraction" ], "authors": [ { "affiliation": "University of Central Florida", "fullName": "Alireza Zaeemzadeh", "givenName": "Alireza", "surname": "Zaeemzadeh", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Trento", "fullName": "Niccolò Bisagno", "givenName": "Niccolò", "surname": "Bisagno", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Trento", "fullName": "Zeno Sambugaro", "givenName": "Zeno", "surname": "Sambugaro", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Trento", "fullName": "Nicola Conci", "givenName": "Nicola", "surname": "Conci", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Central Florida", "fullName": "Nazanin Rahnavard", "givenName": "Nazanin", "surname": "Rahnavard", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Central Florida", "fullName": "Mubarak Shah", "givenName": "Mubarak", "surname": "Shah", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-06-01T00:00:00", "pubType": "proceedings", "pages": "9447-9456", "year": "2021", "issn": null, "isbn": "978-1-6654-4509-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [ { "id": "1yeKwvEPUpa", "name": "pcvpr202145090-09577372s1-mm_450900j447.zip", "size": "883 kB", "location": "https://www.computer.org/csdl/api/v1/extra/pcvpr202145090-09577372s1-mm_450900j447.zip", "__typename": "WebExtraType" } ], "adjacentArticles": { "previous": { "fno": "450900j437", "articleId": "1yeLUakXCEg", "__typename": "AdjacentArticleType" }, "next": { "fno": "450900j457", "articleId": "1yeJpnotqJW", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2021/2812/0/281200b133", "title": "CODEs: Chamfer Out-of-Distribution Examples against Overconfidence Issue", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200b133/1BmFaDYHtgA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200i281", "title": "Semantically Coherent Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200i281/1BmFq1tAb16", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2022/8739/0/873900a163", "title": "RODD: A Self-Supervised Approach for Robust Out-of-Distribution Detection", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2022/873900a163/1G56R8MQAwg", "parentPublication": { "id": "proceedings/cvprw/2022/8739/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600e911", "title": "ViM: Out-Of-Distribution with Virtual-logit Matching", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600e911/1H0LoDIYyB2", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/5555/01/09987694", "title": "Revealing the Distributional Vulnerability of Discriminators by Implicit Generators", "doi": null, "abstractUrl": "/journal/tp/5555/01/09987694/1J7RLU1AUAU", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2023/9346/0/934600f520", "title": "Mixture Outlier Exposure: Towards Out-of-Distribution Detection in Fine-grained Environments", "doi": null, "abstractUrl": "/proceedings-article/wacv/2023/934600f520/1KxUKOwyJJS", "parentPublication": { "id": "proceedings/wacv/2023/9346/0", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bigcomp/2023/7578/0/757800a195", "title": "Attention Masking for Improved Near Out-of-Distribution Image Detection", "doi": null, "abstractUrl": "/proceedings-article/bigcomp/2023/757800a195/1LFLD4mzJ9m", "parentPublication": { "id": "proceedings/bigcomp/2023/7578/0", "title": "2023 IEEE International Conference on Big Data and Smart Computing (BigComp)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300j517", "title": "Unsupervised Out-of-Distribution Detection by Maximum Classifier Discrepancy", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300j517/1hVl9mSRtfO", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2021/8808/0/09412612", "title": "Boundary Optimised Samples Training for Detecting Out-of-Distribution Images", "doi": null, "abstractUrl": "/proceedings-article/icpr/2021/09412612/1tmiAPeMOOs", "parentPublication": { "id": "proceedings/icpr/2021/8808/0", "title": "2020 25th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2021/4899/0/489900d285", "title": "Sample-free white-box out-of-distribution detection for deep learning", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2021/489900d285/1yXsVNgnUA0", "parentPublication": { "id": "proceedings/cvprw/2021/4899/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNz2TCuR", "title": "2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2013", "__typename": "ProceedingType" }, "article": { "id": "12OmNA0MYZM", "doi": "10.1109/CVPR.2013.41", "title": "Wide-Baseline Hair Capture Using Strand-Based Refinement", "normalizedTitle": "Wide-Baseline Hair Capture Using Strand-Based Refinement", "abstract": "We propose a novel algorithm to reconstruct the 3D geometry of human hairs in wide-baseline setups using strand-based refinement. The hair strands are first extracted in each 2D view, and projected onto the 3D visual hull for initialization. The 3D positions of these strands are then refined by optimizing an objective function that takes into account cross-view hair orientation consistency, the visual hull constraint and smoothness constraints defined at the strand, wisp and global levels. Based on the refined strands, the algorithm can reconstruct an approximate hair surface: experiments with synthetic hair models achieve an accuracy of ~3mm. We also show real-world examples to demonstrate the capability to capture full-head hair styles as well as hair in motion with as few as 8 cameras.", "abstracts": [ { "abstractType": "Regular", "content": "We propose a novel algorithm to reconstruct the 3D geometry of human hairs in wide-baseline setups using strand-based refinement. The hair strands are first extracted in each 2D view, and projected onto the 3D visual hull for initialization. The 3D positions of these strands are then refined by optimizing an objective function that takes into account cross-view hair orientation consistency, the visual hull constraint and smoothness constraints defined at the strand, wisp and global levels. Based on the refined strands, the algorithm can reconstruct an approximate hair surface: experiments with synthetic hair models achieve an accuracy of ~3mm. We also show real-world examples to demonstrate the capability to capture full-head hair styles as well as hair in motion with as few as 8 cameras.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose a novel algorithm to reconstruct the 3D geometry of human hairs in wide-baseline setups using strand-based refinement. The hair strands are first extracted in each 2D view, and projected onto the 3D visual hull for initialization. The 3D positions of these strands are then refined by optimizing an objective function that takes into account cross-view hair orientation consistency, the visual hull constraint and smoothness constraints defined at the strand, wisp and global levels. Based on the refined strands, the algorithm can reconstruct an approximate hair surface: experiments with synthetic hair models achieve an accuracy of ~3mm. We also show real-world examples to demonstrate the capability to capture full-head hair styles as well as hair in motion with as few as 8 cameras.", "fno": "4989a265", "keywords": [ "Visual Hull Refinement", "Multi View Stereo", "Hair Reconstruction", "Wide Baseline" ], "authors": [ { "affiliation": null, "fullName": "Linjie Luo", "givenName": "Linjie", "surname": "Luo", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Cha Zhang", "givenName": "Cha", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zhengyou Zhang", "givenName": "Zhengyou", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Szymon Rusinkiewicz", "givenName": "Szymon", "surname": "Rusinkiewicz", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2013-06-01T00:00:00", "pubType": "proceedings", "pages": "265-272", "year": "2013", "issn": "1063-6919", "isbn": "978-0-7695-4989-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4989a257", "articleId": "12OmNy50gdt", "__typename": "AdjacentArticleType" }, "next": { "fno": "4989a273", "articleId": "12OmNzlly0Q", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/smi/2007/2815/0/28150033", "title": "Realistic Hair from a Sketch", "doi": null, "abstractUrl": "/proceedings-article/smi/2007/28150033/12OmNAoUTsL", "parentPublication": { "id": "proceedings/smi/2007/2815/0", "title": "Shape Modeling and Applications, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2013/2322/0/2322a225", "title": "Level-of-Detail Modeling with Artist-Defined Constraints for Photorealistic Hair Rendering", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2013/2322a225/12OmNCcKQNx", "parentPublication": { "id": "proceedings/icvrv/2013/2322/0", "title": "2013 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/simultech/2014/060/0/07095029", "title": "2D hair strands generation based on template matching", "doi": null, "abstractUrl": "/proceedings-article/simultech/2014/07095029/12OmNx5GU7n", "parentPublication": { "id": "proceedings/simultech/2014/060/0", "title": "2014 International Conference on Simulation and Modeling Methodologies, Technologies and Applications (SIMULTECH)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/03/06910280", "title": "2.5D Cartoon Hair Modeling and Manipulation", "doi": null, "abstractUrl": "/journal/tg/2015/03/06910280/13rRUIJuxpC", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07448467", "title": "Adaptive Skinning for Interactive Hair-Solid Simulation", "doi": null, "abstractUrl": "/journal/tg/2017/07/07448467/13rRUygBw7e", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tb/2008/04/ttb2008040484", "title": "Improving Strand Pairing Prediction through Exploring Folding Cooperativity", "doi": null, "abstractUrl": "/journal/tb/2008/04/ttb2008040484/13rRUygBwgl", "parentPublication": { "id": "trans/tb", "title": "IEEE/ACM Transactions on Computational Biology and Bioinformatics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600g133", "title": "HVH: Learning a Hybrid Neural Volumetric Representation for Dynamic Hair Performance Capture", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600g133/1H0OvjmcWUE", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600b516", "title": "NeuralHDHair: Automatic High-fidelity Hair Modeling from a Single Image Using Implicit Neural Representations", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600b516/1H1lkq5sTPq", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2019/3293/0/329300a155", "title": "Strand-Accurate Multi-View Hair Capture", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2019/329300a155/1gyrVg95GSs", "parentPublication": { "id": "proceedings/cvpr/2019/3293/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2019/4752/0/09212824", "title": "Automatic Hair Modeling from One Image", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2019/09212824/1nHRUrDMgE0", "parentPublication": { "id": "proceedings/icvrv/2019/4752/0", "title": "2019 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBEGYFr", "title": "Computer Animation", "acronym": "ca", "groupId": "1000121", "volume": "0", "displayVolume": "0", "year": "1999", "__typename": "ProceedingType" }, "article": { "id": "12OmNCbkQC8", "doi": "10.1109/CA.1999.781199", "title": "Visible Volume Buffer for Efficient Hair Expression and Shadow Generation", "normalizedTitle": "Visible Volume Buffer for Efficient Hair Expression and Shadow Generation", "abstract": "Many researches have been conducted in hair modeling and hair rendering with considerable success. However, the immense number of hair strands present means that memory and CPU time requirements are very severe. To reduce the memory and the time needed for hair modeling and rendering, a visible volume buffer is proposed. Instead of using thousands of thin hairs, the memory usage and hair modeling time can be reduced by using coarse background hairs and fine surface hairs. The background hairs can be constructed by using thick hairs. To improve the look of the hair model, the background hairs near the surface is broken down into numerous thin hairs and rendered. The visible volume buffer is used to determine the surface hairs. The rendering time of the background and surface hairs is found to be faster than conventional hair model by a factor of more than four with little lost in image quality. The visible volume buffer is also used to produce shadow for the hair model.", "abstracts": [ { "abstractType": "Regular", "content": "Many researches have been conducted in hair modeling and hair rendering with considerable success. However, the immense number of hair strands present means that memory and CPU time requirements are very severe. To reduce the memory and the time needed for hair modeling and rendering, a visible volume buffer is proposed. Instead of using thousands of thin hairs, the memory usage and hair modeling time can be reduced by using coarse background hairs and fine surface hairs. The background hairs can be constructed by using thick hairs. To improve the look of the hair model, the background hairs near the surface is broken down into numerous thin hairs and rendered. The visible volume buffer is used to determine the surface hairs. The rendering time of the background and surface hairs is found to be faster than conventional hair model by a factor of more than four with little lost in image quality. The visible volume buffer is also used to produce shadow for the hair model.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Many researches have been conducted in hair modeling and hair rendering with considerable success. However, the immense number of hair strands present means that memory and CPU time requirements are very severe. To reduce the memory and the time needed for hair modeling and rendering, a visible volume buffer is proposed. Instead of using thousands of thin hairs, the memory usage and hair modeling time can be reduced by using coarse background hairs and fine surface hairs. The background hairs can be constructed by using thick hairs. To improve the look of the hair model, the background hairs near the surface is broken down into numerous thin hairs and rendered. The visible volume buffer is used to determine the surface hairs. The rendering time of the background and surface hairs is found to be faster than conventional hair model by a factor of more than four with little lost in image quality. The visible volume buffer is also used to produce shadow for the hair model.", "fno": "01670058", "keywords": [], "authors": [ { "affiliation": "Tokyo Institute of Technology", "fullName": "Waiming Kong", "givenName": "Waiming", "surname": "Kong", "__typename": "ArticleAuthorType" }, { "affiliation": "Tokyo Institute of Technology", "fullName": "Masayuki Nakajima", "givenName": "Masayuki", "surname": "Nakajima", "__typename": "ArticleAuthorType" } ], "idPrefix": "ca", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1999-05-01T00:00:00", "pubType": "proceedings", "pages": "58", "year": "1999", "issn": "1087-4844", "isbn": "0-7695-0167-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "01670210", "articleId": "12OmNzRZpUr", "__typename": "AdjacentArticleType" }, "next": { "fno": "01670070", "articleId": "12OmNyVes12", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNzcxYUX", "title": "2014 International Conference on Simulation and Modeling Methodologies, Technologies and Applications (SIMULTECH)", "acronym": "simultech", "groupId": "1806465", "volume": "0", "displayVolume": "0", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNx5GU7n", "doi": "10.5220/0005107102610266", "title": "2D hair strands generation based on template matching", "normalizedTitle": "2D hair strands generation based on template matching", "abstract": "Hair modelling is an important part of many applications in computer graphics. Since 2D hair strands represent the information of the hair shape and the feature of the hairstyles, the generation of 2D hair strands is an essential part for image-based hair modelling. In this paper, we present a novel algorithm to generate 2D hair strands based on a template matching method. The method first divides a real hairstyle input image into sub-images with the predefined size. For each sub-image, an orientation map is estimated using Gabor filter and the orientation feature is presented by the orientation histogram. Then it matches the orientation histograms between each sub-image and template images in our database. Based on the matching results, the sub-images are replaced by the corresponding manual stroke images to give a clear representation of 2D hair strands. The result is refined by connecting the strands between adjacent sub-images. Finally, based on the control points defined on the 2D hair strands, the spline representation is applied to obtain smooth hair strands. Experimental results indicate that our algorithm is feasible.", "abstracts": [ { "abstractType": "Regular", "content": "Hair modelling is an important part of many applications in computer graphics. Since 2D hair strands represent the information of the hair shape and the feature of the hairstyles, the generation of 2D hair strands is an essential part for image-based hair modelling. In this paper, we present a novel algorithm to generate 2D hair strands based on a template matching method. The method first divides a real hairstyle input image into sub-images with the predefined size. For each sub-image, an orientation map is estimated using Gabor filter and the orientation feature is presented by the orientation histogram. Then it matches the orientation histograms between each sub-image and template images in our database. Based on the matching results, the sub-images are replaced by the corresponding manual stroke images to give a clear representation of 2D hair strands. The result is refined by connecting the strands between adjacent sub-images. Finally, based on the control points defined on the 2D hair strands, the spline representation is applied to obtain smooth hair strands. Experimental results indicate that our algorithm is feasible.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Hair modelling is an important part of many applications in computer graphics. Since 2D hair strands represent the information of the hair shape and the feature of the hairstyles, the generation of 2D hair strands is an essential part for image-based hair modelling. In this paper, we present a novel algorithm to generate 2D hair strands based on a template matching method. The method first divides a real hairstyle input image into sub-images with the predefined size. For each sub-image, an orientation map is estimated using Gabor filter and the orientation feature is presented by the orientation histogram. Then it matches the orientation histograms between each sub-image and template images in our database. Based on the matching results, the sub-images are replaced by the corresponding manual stroke images to give a clear representation of 2D hair strands. The result is refined by connecting the strands between adjacent sub-images. Finally, based on the control points defined on the 2D hair strands, the spline representation is applied to obtain smooth hair strands. Experimental results indicate that our algorithm is feasible.", "fno": "07095029", "keywords": [ "Hair", "Manuals", "Histograms", "Splines Mathematics", "Gabor Filters", "Geometry", "Computational Modeling", "Spline Representation", "Hair Strands Generation", "Orientation Map", "Template Matching" ], "authors": [ { "affiliation": "School of Electrical Engineering and Computer Science, University of Ottawa, 75 Laurier Avenue East, Canada", "fullName": "Chao Sun", "givenName": "Chao", "surname": "Sun", "__typename": "ArticleAuthorType" }, { "affiliation": "School of Electrical Engineering and Computer Science, University of Ottawa, 75 Laurier Avenue East, Canada", "fullName": "Fatemeh Cheraghchi", "givenName": "Fatemeh", "surname": "Cheraghchi", "__typename": "ArticleAuthorType" }, { "affiliation": "School of Electrical Engineering and Computer Science, University of Ottawa, 75 Laurier Avenue East, Canada", "fullName": "Won-Sook Lee", "givenName": "Won-Sook", "surname": "Lee", "__typename": "ArticleAuthorType" } ], "idPrefix": "simultech", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-08-01T00:00:00", "pubType": "proceedings", "pages": "261-266", "year": "2014", "issn": null, "isbn": "978-989-758-060-4", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07095028", "articleId": "12OmNC8Msr5", "__typename": "AdjacentArticleType" }, "next": { "fno": "07095030", "articleId": "12OmNvjQ8S8", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2013/4989/0/4989a265", "title": "Wide-Baseline Hair Capture Using Strand-Based Refinement", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2013/4989a265/12OmNA0MYZM", "parentPublication": { "id": "proceedings/cvpr/2013/4989/0", "title": "2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2000/0743/0/07430303", "title": "Automatic Generation of Hair Texture with Line Integral Convolution", "doi": null, "abstractUrl": "/proceedings-article/iv/2000/07430303/12OmNAYoKl4", "parentPublication": { "id": "proceedings/iv/2000/0743/0", "title": "2000 IEEE Conference on Information Visualization. An International Conference on Computer Visualization and Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2013/2246/0/2246a318", "title": "Skeleton-Based Anime Hair Modeling and Visualization", "doi": null, "abstractUrl": "/proceedings-article/cw/2013/2246a318/12OmNykCceO", "parentPublication": { "id": "proceedings/cw/2013/2246/0", "title": "2013 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2015/9403/0/9403a306", "title": "Modeling Curly Hair Based on Static Super-Helices", "doi": null, "abstractUrl": "/proceedings-article/cw/2015/9403a306/12OmNznkK1a", "parentPublication": { "id": "proceedings/cw/2015/9403/0", "title": "2015 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/03/06910280", "title": "2.5D Cartoon Hair Modeling and Manipulation", "doi": null, "abstractUrl": "/journal/tg/2015/03/06910280/13rRUIJuxpC", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2006/07/i1025", "title": "A Generative Sketch Model for Human Hair Analysis and Synthesis", "doi": null, "abstractUrl": "/journal/tp/2006/07/i1025/13rRUwdrdLQ", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07448467", "title": "Adaptive Skinning for Interactive Hair-Solid Simulation", "doi": null, "abstractUrl": "/journal/tg/2017/07/07448467/13rRUygBw7e", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600b516", "title": "NeuralHDHair: Automatic High-fidelity Hair Modeling from a Single Image Using Implicit Neural Representations", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600b516/1H1lkq5sTPq", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/07/08964443", "title": "DeepSketchHair: Deep Sketch-Based 3D Hair Modeling", "doi": null, "abstractUrl": "/journal/tg/2021/07/08964443/1gLZSnCp3Ko", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icvrv/2019/4752/0/09212824", "title": "Automatic Hair Modeling from One Image", "doi": null, "abstractUrl": "/proceedings-article/icvrv/2019/09212824/1nHRUrDMgE0", "parentPublication": { "id": "proceedings/icvrv/2019/4752/0", "title": "2019 International Conference on Virtual Reality and Visualization (ICVRV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNrF2DIa", "title": "2017 21st International Conference Information Visualisation (IV)", "acronym": "iv", "groupId": "1000370", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNB7LvBm", "doi": "10.1109/iV.2017.56", "title": "Visual Analytics for Electronic Intelligence: Challenges and Opportunities", "normalizedTitle": "Visual Analytics for Electronic Intelligence: Challenges and Opportunities", "abstract": "In this paper, we present the field of Electronic Intelligence (ELINT) and the issues that it raises for visual analytics. ELINT aggregates many of the actual issues that visual analytics face such as huge amounts of data, complex data, complex tasks, missing data and unreliable data. This aggregation of specificities makes ELINT a domain raising many visualization issues. This paper identifies the challenges of ELINT by describing its specificities and identifies the opportunities that ELINT raises for the visualization domain.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, we present the field of Electronic Intelligence (ELINT) and the issues that it raises for visual analytics. ELINT aggregates many of the actual issues that visual analytics face such as huge amounts of data, complex data, complex tasks, missing data and unreliable data. This aggregation of specificities makes ELINT a domain raising many visualization issues. This paper identifies the challenges of ELINT by describing its specificities and identifies the opportunities that ELINT raises for the visualization domain.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, we present the field of Electronic Intelligence (ELINT) and the issues that it raises for visual analytics. ELINT aggregates many of the actual issues that visual analytics face such as huge amounts of data, complex data, complex tasks, missing data and unreliable data. This aggregation of specificities makes ELINT a domain raising many visualization issues. This paper identifies the challenges of ELINT by describing its specificities and identifies the opportunities that ELINT raises for the visualization domain.", "fno": "0831a422", "keywords": [ "Data Analysis", "Data Visualisation", "Electronic Intelligence", "ELINT Aggregates", "Visual Analytics", "Visualization Domain", "Visualization Issues", "Unreliable Data", "Complex Tasks", "Complex Data", "Radar", "Time Frequency Analysis", "Sensors", "Visual Analytics", "Data Visualization", "Complexity Theory", "Visual Analytic Application", "Electronic Intelligence", "Multidimensional Data", "Temporal Data" ], "authors": [ { "affiliation": null, "fullName": "Cantu Alma", "givenName": "Cantu", "surname": "Alma", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Grisvard Olivier", "givenName": "Grisvard", "surname": "Olivier", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Duval Thierry", "givenName": "Duval", "surname": "Thierry", "__typename": "ArticleAuthorType" } ], "idPrefix": "iv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-07-01T00:00:00", "pubType": "proceedings", "pages": "422-426", "year": "2017", "issn": "2375-0138", "isbn": "978-1-5386-0831-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0831a416", "articleId": "12OmNzVoBMz", "__typename": "AdjacentArticleType" }, "next": { "fno": "0831a427", "articleId": "12OmNzlD9AL", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/bdva/2015/7343/0/07314288", "title": "Challenges and Perspectives in Big Eye-Movement Data Visual Analytics", "doi": null, "abstractUrl": "/proceedings-article/bdva/2015/07314288/12OmNButq1p", "parentPublication": { "id": "proceedings/bdva/2015/7343/0", "title": "2015 Big Data Visual Analytics (BDVA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icse-c/2016/4205/0/4205a902", "title": "Software Analytics: Challenges and Opportunities", "doi": null, "abstractUrl": "/proceedings-article/icse-c/2016/4205a902/12OmNCesrcI", "parentPublication": { "id": "proceedings/icse-c/2016/4205/0", "title": "2016 IEEE/ACM 38th International Conference on Software Engineering Companion (ICSE-C)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bdva/2015/7343/0/07314296", "title": "Immersive Analytics", "doi": null, "abstractUrl": "/proceedings-article/bdva/2015/07314296/12OmNzVXNSO", "parentPublication": { "id": "proceedings/bdva/2015/7343/0", "title": "2015 Big Data Visual Analytics (BDVA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2008/01/mcg2008010018", "title": "An Information-Theoretic View of Visual Analytics", "doi": null, "abstractUrl": "/magazine/cg/2008/01/mcg2008010018/13rRUB6SpRW", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2011/05/mcg2011050018", "title": "Graph Analytics—Lessons Learned and Challenges Ahead", "doi": null, "abstractUrl": "/magazine/cg/2011/05/mcg2011050018/13rRUxASu6j", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2012/04/mcg2012040063", "title": "The Top 10 Challenges in Extreme-Scale Visual Analytics", "doi": null, "abstractUrl": "/magazine/cg/2012/04/mcg2012040063/13rRUxC0SGA", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/co/2013/07/mco2013070030", "title": "Visual Analytics Support for Intelligence Analysis", "doi": null, "abstractUrl": "/magazine/co/2013/07/mco2013070030/13rRUxD9h0P", "parentPublication": { "id": "mags/co", "title": "Computer", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2018/01/08023780", "title": "EVA: Visual Analytics to Identify Fraudulent Events", "doi": null, "abstractUrl": "/journal/tg/2018/01/08023780/13rRUxNEqQ0", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2015/04/mcg2015040052", "title": "Design Challenges and Opportunities for Eco-Feedback in the Home", "doi": null, "abstractUrl": "/magazine/cg/2015/04/mcg2015040052/13rRUxYINaN", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/percom-workshops/2021/0424/0/09431071", "title": "Keynote: Edge Intelligence - Emerging Solutions and Open Challenges", "doi": null, "abstractUrl": "/proceedings-article/percom-workshops/2021/09431071/1tROOh6SQWk", "parentPublication": { "id": "proceedings/percom-workshops/2021/0424/0", "title": "2021 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBzRNrw", "title": "2013 46th Hawaii International Conference on System Sciences", "acronym": "hicss", "groupId": "1000730", "volume": "0", "displayVolume": "0", "year": "2013", "__typename": "ProceedingType" }, "article": { "id": "12OmNqJ8tq4", "doi": "10.1109/HICSS.2013.58", "title": "A Role for Reasoning in Visual Analytics", "normalizedTitle": "A Role for Reasoning in Visual Analytics", "abstract": "Analysis supported by interactive visual interfaces is a complex process. It involves computational analytics (the visualization of both raw and derived data) and an analytical process which requires a human to extract knowledge from the data by directly interacting and manipulating both the visual and analytical components of the system. These two types of analytics are complementary and the goal of this paper is to understand interplay between the two. In this paper we discuss how a study of human reasoning and reasoning-supported cognitive processes complement the current emphasis on computational analysis and visualization. We define this process as reasoning analytics and present mechanisms by which this process may be studied.", "abstracts": [ { "abstractType": "Regular", "content": "Analysis supported by interactive visual interfaces is a complex process. It involves computational analytics (the visualization of both raw and derived data) and an analytical process which requires a human to extract knowledge from the data by directly interacting and manipulating both the visual and analytical components of the system. These two types of analytics are complementary and the goal of this paper is to understand interplay between the two. In this paper we discuss how a study of human reasoning and reasoning-supported cognitive processes complement the current emphasis on computational analysis and visualization. We define this process as reasoning analytics and present mechanisms by which this process may be studied.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Analysis supported by interactive visual interfaces is a complex process. It involves computational analytics (the visualization of both raw and derived data) and an analytical process which requires a human to extract knowledge from the data by directly interacting and manipulating both the visual and analytical components of the system. These two types of analytics are complementary and the goal of this paper is to understand interplay between the two. In this paper we discuss how a study of human reasoning and reasoning-supported cognitive processes complement the current emphasis on computational analysis and visualization. We define this process as reasoning analytics and present mechanisms by which this process may be studied.", "fno": "4892b495", "keywords": [ "Cognition", "Data Visualization", "Decision Making", "Visual Analytics", "Problem Solving", "Visual Analytics", "Cognitive Systems", "Empirical Studies" ], "authors": [ { "affiliation": null, "fullName": "Tera Marie Green", "givenName": "Tera Marie", "surname": "Green", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ross Maciejewski", "givenName": "Ross", "surname": "Maciejewski", "__typename": "ArticleAuthorType" } ], "idPrefix": "hicss", "isOpenAccess": true, "showRecommendedArticles": true, "showBuyMe": false, "hasPdf": true, "pubDate": "2013-01-01T00:00:00", "pubType": "proceedings", "pages": "1495-1504", "year": "2013", "issn": "1530-1605", "isbn": "978-1-4673-5933-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4892b485", "articleId": "12OmNCgrD1q", "__typename": "AdjacentArticleType" }, "next": { "fno": "4892b505", "articleId": "12OmNAolGTi", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/hicss/2013/4892/0/4892c416", "title": "Visual Analytics for Public Health: Supporting Knowledge Construction and Decision-Making", "doi": null, "abstractUrl": "/proceedings-article/hicss/2013/4892c416/12OmNrJiCNq", "parentPublication": { "id": "proceedings/hicss/2013/4892/0", "title": "2013 46th Hawaii International Conference on System Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/hicss/2011/9618/0/05718616", "title": "Pair Analytics: Capturing Reasoning Processes in Collaborative Visual Analytics", "doi": null, "abstractUrl": "/proceedings-article/hicss/2011/05718616/12OmNvAiShB", "parentPublication": { "id": "proceedings/hicss/2011/9618/0", "title": "2011 44th Hawaii International Conference on System Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2008/01/mcg2008010018", "title": "An Information-Theoretic View of Visual Analytics", "doi": null, "abstractUrl": "/magazine/cg/2008/01/mcg2008010018/13rRUB6SpRW", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2012/12/ttg2012122908", "title": "The User Puzzle—Explaining the Interaction with Visual Analytics Systems", "doi": null, "abstractUrl": "/journal/tg/2012/12/ttg2012122908/13rRUIIVlcH", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2014/04/mcg2014040008", "title": "Semantic Interaction for Visual Analytics: Toward Coupling Cognition and Computation", "doi": null, "abstractUrl": "/magazine/cg/2014/04/mcg2014040008/13rRUwwslv3", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/01/07534883", "title": "Characterizing Guidance in Visual Analytics", "doi": null, "abstractUrl": "/journal/tg/2017/01/07534883/13rRUxBa568", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2013/07/ttg2013071076", "title": "Guest Editors' Introduction: Special Section on the IEEE Conference on Visual Analytics Science and Technology (VAST)", "doi": null, "abstractUrl": "/journal/tg/2013/07/ttg2013071076/13rRUxOdD2D", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2016/01/07192716", "title": "The Role of Uncertainty, Awareness, and Trust in Visual Analytics", "doi": null, "abstractUrl": "/journal/tg/2016/01/07192716/13rRUxlgxTo", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/co/2013/07/mco2013070020", "title": "Visual Analytics: Seeking the Unknown", "doi": null, "abstractUrl": "/magazine/co/2013/07/mco2013070020/13rRUy0HYNj", "parentPublication": { "id": "mags/co", "title": "Computer", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iv/2021/3827/0/382700a211", "title": "Visual Analytics and Similarity Search - Interest-based Similarity Search in Scientific Data", "doi": null, "abstractUrl": "/proceedings-article/iv/2021/382700a211/1y4oJ30dpcY", "parentPublication": { "id": "proceedings/iv/2021/3827/0", "title": "2021 25th International Conference Information Visualisation (IV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAR1b0Z", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "acronym": "cvprw", "groupId": "1001809", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNAndigF", "doi": "10.1109/CVPRW.2017.169", "title": "Generating 5D Light Fields in Scattering Media for Representing 3D Images", "normalizedTitle": "Generating 5D Light Fields in Scattering Media for Representing 3D Images", "abstract": "In this paper, we propose a novel method for displaying 3D images based on a 5D light field representation. In our method, the light fields emitted by a light field projector are projected into 3D scattering media such as fog. The intensity of light lays projected into the scattering media decreases because of the scattering effect of the media. As a result, 5D light fields are generated in the scattering media. The proposed method models the relationship between the 5D light fields and observed images, and uses the relationship for projecting light fields so that the observed image changes according to the viewpoint of observers. In order to achieve accurate and efficient 3D image representation, we describe the relationship not by using a parametric model, but by using an observation based model obtained from a point spread function (PSF) of scattering media. The experimental results show the efficiency of the proposed method.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper, we propose a novel method for displaying 3D images based on a 5D light field representation. In our method, the light fields emitted by a light field projector are projected into 3D scattering media such as fog. The intensity of light lays projected into the scattering media decreases because of the scattering effect of the media. As a result, 5D light fields are generated in the scattering media. The proposed method models the relationship between the 5D light fields and observed images, and uses the relationship for projecting light fields so that the observed image changes according to the viewpoint of observers. In order to achieve accurate and efficient 3D image representation, we describe the relationship not by using a parametric model, but by using an observation based model obtained from a point spread function (PSF) of scattering media. The experimental results show the efficiency of the proposed method.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper, we propose a novel method for displaying 3D images based on a 5D light field representation. In our method, the light fields emitted by a light field projector are projected into 3D scattering media such as fog. The intensity of light lays projected into the scattering media decreases because of the scattering effect of the media. As a result, 5D light fields are generated in the scattering media. The proposed method models the relationship between the 5D light fields and observed images, and uses the relationship for projecting light fields so that the observed image changes according to the viewpoint of observers. In order to achieve accurate and efficient 3D image representation, we describe the relationship not by using a parametric model, but by using an observation based model obtained from a point spread function (PSF) of scattering media. The experimental results show the efficiency of the proposed method.", "fno": "0733b287", "keywords": [ "Scattering", "Media", "Three Dimensional Displays", "Observers", "Visualization", "Two Dimensional Displays", "Parametric Statistics" ], "authors": [ { "affiliation": null, "fullName": "Eri Yuasa", "givenName": "Eri", "surname": "Yuasa", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Fumihiko Sakaue", "givenName": "Fumihiko", "surname": "Sakaue", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jun Sato", "givenName": "Jun", "surname": "Sato", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvprw", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-07-01T00:00:00", "pubType": "proceedings", "pages": "1287-1294", "year": "2017", "issn": "2160-7516", "isbn": "978-1-5386-0733-6", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0733b277", "articleId": "12OmNs0C9Cb", "__typename": "AdjacentArticleType" }, "next": { "fno": "0733b295", "articleId": "12OmNrkT7ID", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2016/8851/0/8851b745", "title": "Heterogeneous Light Fields", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2016/8851b745/12OmNA0dMG8", "parentPublication": { "id": "proceedings/cvpr/2016/8851/0", "title": "2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d415", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d415/12OmNBV9Igo", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209e382", "title": "Light Transport Refocusing for Unknown Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209e382/12OmNqzu6Nb", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cis/2012/4896/0/4896a396", "title": "Fast Multiple Scattering in Participating Media with Beamlet Decomposition", "doi": null, "abstractUrl": "/proceedings-article/cis/2012/4896a396/12OmNwekjJa", "parentPublication": { "id": "proceedings/cis/2012/4896/0", "title": "2012 Eighth International Conference on Computational Intelligence and Security", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sitis/2013/3211/0/3211a093", "title": "Multiple-Scattering Optical Tomography with Layered Material", "doi": null, "abstractUrl": "/proceedings-article/sitis/2013/3211a093/12OmNzRZpYR", "parentPublication": { "id": "proceedings/sitis/2013/3211/0", "title": "2013 International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2013/03/mcg2013030066", "title": "Double- and Multiple-Scattering Effects in Translucent Materials", "doi": null, "abstractUrl": "/magazine/cg/2013/03/mcg2013030066/13rRUIJcWfX", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07452672", "title": "Expressive Single Scattering for Light Shaft Stylization", "doi": null, "abstractUrl": "/journal/tg/2017/07/07452672/13rRUx0xPZB", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2017/09/07577857", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/journal/tp/2017/09/07577857/13rRUxYIMWv", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/07/08600345", "title": "Precomputed Multiple Scattering for Rapid Light Simulation in Participating Media", "doi": null, "abstractUrl": "/journal/tg/2020/07/08600345/17D45Xh13tH", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09715049", "title": "Collimated Whole Volume Light Scattering in Homogeneous Finite Media", "doi": null, "abstractUrl": "/journal/tg/5555/01/09715049/1B2DbhImWwE", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNrNh0vw", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "acronym": "icpr", "groupId": "1000545", "volume": "0", "displayVolume": "0", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNqzu6Nb", "doi": "10.1109/ICPR.2014.750", "title": "Light Transport Refocusing for Unknown Scattering Medium", "normalizedTitle": "Light Transport Refocusing for Unknown Scattering Medium", "abstract": "In this paper we propose a new light transport refocusing method for depth estimation as well as for investigation inside scattering media with unknown scattering properties. Propagated visible light rays through scattering media are utilized in our proposed refocusing method. We use 2D light source to illuminate the scattering media and 2D image sensor for capturing transported rays. The proposed method that uses 4D light transport can clearly visualize shallow depth, as well as deep depth plane of the medium. We apply our light transport refocusing method for depth estimation using conventional depth-from-focus method and for clear visualization by descattering the light rays passing through the medium. To evaluate the effectiveness we have done experiments using acrylic and milk-water type scattering medium in various optical and geometrical conditions. Finally, we show up the results of depth estimation and clear visualization, as well as with numeric evaluation.", "abstracts": [ { "abstractType": "Regular", "content": "In this paper we propose a new light transport refocusing method for depth estimation as well as for investigation inside scattering media with unknown scattering properties. Propagated visible light rays through scattering media are utilized in our proposed refocusing method. We use 2D light source to illuminate the scattering media and 2D image sensor for capturing transported rays. The proposed method that uses 4D light transport can clearly visualize shallow depth, as well as deep depth plane of the medium. We apply our light transport refocusing method for depth estimation using conventional depth-from-focus method and for clear visualization by descattering the light rays passing through the medium. To evaluate the effectiveness we have done experiments using acrylic and milk-water type scattering medium in various optical and geometrical conditions. Finally, we show up the results of depth estimation and clear visualization, as well as with numeric evaluation.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "In this paper we propose a new light transport refocusing method for depth estimation as well as for investigation inside scattering media with unknown scattering properties. Propagated visible light rays through scattering media are utilized in our proposed refocusing method. We use 2D light source to illuminate the scattering media and 2D image sensor for capturing transported rays. The proposed method that uses 4D light transport can clearly visualize shallow depth, as well as deep depth plane of the medium. We apply our light transport refocusing method for depth estimation using conventional depth-from-focus method and for clear visualization by descattering the light rays passing through the medium. To evaluate the effectiveness we have done experiments using acrylic and milk-water type scattering medium in various optical and geometrical conditions. Finally, we show up the results of depth estimation and clear visualization, as well as with numeric evaluation.", "fno": "5209e382", "keywords": [ "Scattering", "Media", "Cameras", "Light Sources", "Lenses", "Visualization" ], "authors": [ { "affiliation": null, "fullName": "Md. Abdul Mannan", "givenName": "Md. Abdul", "surname": "Mannan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Seiichi Tagawa", "givenName": "Seiichi", "surname": "Tagawa", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Toru Tamaki", "givenName": "Toru", "surname": "Tamaki", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Hajime Nagahara", "givenName": "Hajime", "surname": "Nagahara", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yasuhiro Mukaigawa", "givenName": "Yasuhiro", "surname": "Mukaigawa", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Yasushi Yagi", "givenName": "Yasushi", "surname": "Yagi", "__typename": "ArticleAuthorType" } ], "idPrefix": "icpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-08-01T00:00:00", "pubType": "proceedings", "pages": "4382-4387", "year": "2014", "issn": "1051-4651", "isbn": "978-1-4799-5209-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "5209e376", "articleId": "12OmNCxbXKd", "__typename": "AdjacentArticleType" }, "next": { "fno": "5209e388", "articleId": "12OmNzayNjV", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icme/2014/4761/0/06890175", "title": "Anti-aliasing for light field rendering", "doi": null, "abstractUrl": "/proceedings-article/icme/2014/06890175/12OmNBBQZsp", "parentPublication": { "id": "proceedings/icme/2014/4761/0", "title": "2014 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d415", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d415/12OmNBV9Igo", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2017/1032/0/1032c420", "title": "Depth and Image Restoration from Light Field in a Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/iccv/2017/1032c420/12OmNxjjEm8", "parentPublication": { "id": "proceedings/iccv/2017/1032/0", "title": "2017 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sitis/2013/3211/0/3211a093", "title": "Multiple-Scattering Optical Tomography with Layered Material", "doi": null, "abstractUrl": "/proceedings-article/sitis/2013/3211a093/12OmNzRZpYR", "parentPublication": { "id": "proceedings/sitis/2013/3211/0", "title": "2013 International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2014/5188/0/06831819", "title": "Digital refocusing with incoherent holography", "doi": null, "abstractUrl": "/proceedings-article/iccp/2014/06831819/12OmNzgeLBy", "parentPublication": { "id": "proceedings/iccp/2014/5188/0", "title": "2014 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2013/6463/0/06528300", "title": "Descattering of transmissive observation using Parallel High-Frequency Illumination", "doi": null, "abstractUrl": "/proceedings-article/iccp/2013/06528300/12OmNzmclka", "parentPublication": { "id": "proceedings/iccp/2013/6463/0", "title": "2013 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d595", "title": "Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d595/12OmNzt0INA", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/02/v0342", "title": "Light Scattering from Filaments", "doi": null, "abstractUrl": "/journal/tg/2007/02/v0342/13rRUwI5TXt", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2017/09/07577857", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/journal/tp/2017/09/07577857/13rRUxYIMWv", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09715049", "title": "Collimated Whole Volume Light Scattering in Homogeneous Finite Media", "doi": null, "abstractUrl": "/journal/tg/5555/01/09715049/1B2DbhImWwE", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxE2mWp", "title": "2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNxWLTlm", "doi": "10.1109/CVPR.2010.5540216", "title": "Analysis of light transport in scattering media", "normalizedTitle": "Analysis of light transport in scattering media", "abstract": "We propose a new method to analyze light transport in homogeneous scattering media. The incident light undergoes multiple bounces in translucent objects, and produces a complex light field. Our method analyzes the light transport in two steps. First, single and multiple scattering are separated by projecting high-frequency stripe patterns. Then, multiple scattering is decomposed into each bounce component based on the light transport equation. The light field for each bounce is recursively estimated. Experimental results show that light transport in scattering media can be decomposed and visualized for each bounce.", "abstracts": [ { "abstractType": "Regular", "content": "We propose a new method to analyze light transport in homogeneous scattering media. The incident light undergoes multiple bounces in translucent objects, and produces a complex light field. Our method analyzes the light transport in two steps. First, single and multiple scattering are separated by projecting high-frequency stripe patterns. Then, multiple scattering is decomposed into each bounce component based on the light transport equation. The light field for each bounce is recursively estimated. Experimental results show that light transport in scattering media can be decomposed and visualized for each bounce.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose a new method to analyze light transport in homogeneous scattering media. The incident light undergoes multiple bounces in translucent objects, and produces a complex light field. Our method analyzes the light transport in two steps. First, single and multiple scattering are separated by projecting high-frequency stripe patterns. Then, multiple scattering is decomposed into each bounce component based on the light transport equation. The light field for each bounce is recursively estimated. Experimental results show that light transport in scattering media can be decomposed and visualized for each bounce.", "fno": "05540216", "keywords": [], "authors": [ { "affiliation": "Osaka University", "fullName": "Yasuhiro Mukaigawa", "givenName": "Yasuhiro", "surname": "Mukaigawa", "__typename": "ArticleAuthorType" }, { "affiliation": "Osaka University", "fullName": "Yasushi Yagi", "givenName": "Yasushi", "surname": "Yagi", "__typename": "ArticleAuthorType" }, { "affiliation": "MIT Media Lab", "fullName": "Ramesh Raskar", "givenName": "Ramesh", "surname": "Raskar", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-06-01T00:00:00", "pubType": "proceedings", "pages": "153-160", "year": "2010", "issn": null, "isbn": "978-1-4244-6984-0", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "05540219", "articleId": "12OmNARRYlG", "__typename": "AdjacentArticleType" }, "next": { "fno": "05540217", "articleId": "12OmNxw5Buc", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvprw/2017/0733/0/0733b287", "title": "Generating 5D Light Fields in Scattering Media for Representing 3D Images", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2017/0733b287/12OmNAndigF", "parentPublication": { "id": "proceedings/cvprw/2017/0733/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dim/2007/2939/0/29390337", "title": "Light Transport Analysis for 3D Photography", "doi": null, "abstractUrl": "/proceedings-article/3dim/2007/29390337/12OmNqOffAc", "parentPublication": { "id": "proceedings/3dim/2007/2939/0", "title": "2007 6th International Conference on 3-D Digital Imaging and Modeling", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209e382", "title": "Light Transport Refocusing for Unknown Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209e382/12OmNqzu6Nb", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2012/1226/0/047O1A02", "title": "Decomposing global light transport using time of flight imaging", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2012/047O1A02/12OmNwCJOTg", "parentPublication": { "id": "proceedings/cvpr/2012/1226/0", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cis/2012/4896/0/4896a396", "title": "Fast Multiple Scattering in Participating Media with Beamlet Decomposition", "doi": null, "abstractUrl": "/proceedings-article/cis/2012/4896a396/12OmNwekjJa", "parentPublication": { "id": "proceedings/cis/2012/4896/0", "title": "2012 Eighth International Conference on Computational Intelligence and Security", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2018/2526/0/08368461", "title": "Acquiring and characterizing plane-to-ray indirect light transport", "doi": null, "abstractUrl": "/proceedings-article/iccp/2018/08368461/12OmNzkMlWO", "parentPublication": { "id": "proceedings/iccp/2018/2526/0", "title": "2018 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2005/2334/1/23340420", "title": "Structured Light in Scattering Media", "doi": null, "abstractUrl": "/proceedings-article/iccv/2005/23340420/12OmNzvz6Oz", "parentPublication": { "id": "proceedings/iccv/2005/2334/2", "title": "Computer Vision, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/02/v0342", "title": "Light Scattering from Filaments", "doi": null, "abstractUrl": "/journal/tg/2007/02/v0342/13rRUwI5TXt", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/07/08600345", "title": "Precomputed Multiple Scattering for Rapid Light Simulation in Participating Media", "doi": null, "abstractUrl": "/journal/tg/2020/07/08600345/17D45Xh13tH", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2021/04/08877764", "title": "Programmable Non-Epipolar Indirect Light Transport: Capture and Analysis", "doi": null, "abstractUrl": "/journal/tg/2021/04/08877764/1emy95qb1NS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAXxXaK", "title": "2017 IEEE International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNxjjEm8", "doi": "10.1109/ICCV.2017.263", "title": "Depth and Image Restoration from Light Field in a Scattering Medium", "normalizedTitle": "Depth and Image Restoration from Light Field in a Scattering Medium", "abstract": "Traditional imaging methods and computer vision algorithms are often ineffective when images are acquired in scattering media, such as underwater, fog, and biological tissue. Here, we explore the use of light field imaging and algorithms for image restoration and depth estimation that address the image degradation from the medium. Towards this end, we make the following three contributions. First, we present a new single image restoration algorithm which removes backscatter and attenuation from images better than existing methods do, and apply it to each view in the light field. Second, we combine a novel transmission based depth cue with existing correspondence and defocus cues to improve light field depth estimation. In densely scattering media, our transmission depth cue is critical for depth estimation since the images have low signal to noise ratios which significantly degrades the performance of the correspondence and defocus cues. Finally, we propose shearing and refocusing multiple views of the light field to recover a single image of higher quality than what is possible from a single view. We demonstrate the benefits of our method through extensive experimental results in a water tank.", "abstracts": [ { "abstractType": "Regular", "content": "Traditional imaging methods and computer vision algorithms are often ineffective when images are acquired in scattering media, such as underwater, fog, and biological tissue. Here, we explore the use of light field imaging and algorithms for image restoration and depth estimation that address the image degradation from the medium. Towards this end, we make the following three contributions. First, we present a new single image restoration algorithm which removes backscatter and attenuation from images better than existing methods do, and apply it to each view in the light field. Second, we combine a novel transmission based depth cue with existing correspondence and defocus cues to improve light field depth estimation. In densely scattering media, our transmission depth cue is critical for depth estimation since the images have low signal to noise ratios which significantly degrades the performance of the correspondence and defocus cues. Finally, we propose shearing and refocusing multiple views of the light field to recover a single image of higher quality than what is possible from a single view. We demonstrate the benefits of our method through extensive experimental results in a water tank.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Traditional imaging methods and computer vision algorithms are often ineffective when images are acquired in scattering media, such as underwater, fog, and biological tissue. Here, we explore the use of light field imaging and algorithms for image restoration and depth estimation that address the image degradation from the medium. Towards this end, we make the following three contributions. First, we present a new single image restoration algorithm which removes backscatter and attenuation from images better than existing methods do, and apply it to each view in the light field. Second, we combine a novel transmission based depth cue with existing correspondence and defocus cues to improve light field depth estimation. In densely scattering media, our transmission depth cue is critical for depth estimation since the images have low signal to noise ratios which significantly degrades the performance of the correspondence and defocus cues. Finally, we propose shearing and refocusing multiple views of the light field to recover a single image of higher quality than what is possible from a single view. We demonstrate the benefits of our method through extensive experimental results in a water tank.", "fno": "1032c420", "keywords": [ "Estimation Theory", "Image Restoration", "Scattering", "Scattering Medium", "Light Field Imaging", "Transmission Based Depth Cue", "Light Field Depth Estimation", "Densely Scattering Media", "Transmission Depth Cue", "Image Degradation", "Image Restoration Algorithm", "Multiple Views Shearing", "Multiple Views Refocusing", "Scattering", "Image Restoration", "Cameras", "Backscatter", "Estimation", "Media" ], "authors": [ { "affiliation": null, "fullName": "Jiandong Tian", "givenName": "Jiandong", "surname": "Tian", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zak Murez", "givenName": "Zak", "surname": "Murez", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Tong Cui", "givenName": "Tong", "surname": "Cui", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Zhen Zhang", "givenName": "Zhen", "surname": "Zhang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "David Kriegman", "givenName": "David", "surname": "Kriegman", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ravi Ramamoorthi", "givenName": "Ravi", "surname": "Ramamoorthi", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-10-01T00:00:00", "pubType": "proceedings", "pages": "2420-2429", "year": "2017", "issn": "2380-7504", "isbn": "978-1-5386-1032-9", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "1032c410", "articleId": "12OmNB8kHOV", "__typename": "AdjacentArticleType" }, "next": { "fno": "1032c430", "articleId": "12OmNylKAMJ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2009/3992/0/05206855", "title": "(De) focusing on global light transport for active scene recovery", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2009/05206855/12OmNAoDi7C", "parentPublication": { "id": "proceedings/cvpr/2009/3992/0", "title": "2009 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d415", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d415/12OmNBV9Igo", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2013/3022/0/3022a037", "title": "External Mask Based Depth and Light Field Camera", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2013/3022a037/12OmNCctfnA", "parentPublication": { "id": "proceedings/iccvw/2013/3022/0", "title": "2013 IEEE International Conference on Computer Vision Workshops (ICCVW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209e382", "title": "Light Transport Refocusing for Unknown Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209e382/12OmNqzu6Nb", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2013/2840/0/2840a673", "title": "Depth from Combining Defocus and Correspondence Using Light-Field Cameras", "doi": null, "abstractUrl": "/proceedings-article/iccv/2013/2840a673/12OmNxETas6", "parentPublication": { "id": "proceedings/iccv/2013/2840/0", "title": "2013 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ssiai/2018/6568/0/08470347", "title": "Underwater Image Restoration using Deep Networks to Estimate Background Light and Scene Depth", "doi": null, "abstractUrl": "/proceedings-article/ssiai/2018/08470347/13WBGN0lJ0o", "parentPublication": { "id": "proceedings/ssiai/2018/6568/0", "title": "2018 IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2018/3365/0/08446366", "title": "The Depth Light", "doi": null, "abstractUrl": "/proceedings-article/vr/2018/08446366/13bd1AITnaG", "parentPublication": { "id": "proceedings/vr/2018/3365/0", "title": "2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2018/10/08022875", "title": "Robust Light Field Depth Estimation Using Occlusion-Noise Aware Data Costs", "doi": null, "abstractUrl": "/journal/tp/2018/10/08022875/13rRUEgaru8", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2017/09/07577857", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/journal/tp/2017/09/07577857/13rRUxYIMWv", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2017/03/07452621", "title": "Shape Estimation from Shading, Defocus, and Correspondence Using Light-Field Angular Coherence", "doi": null, "abstractUrl": "/journal/tp/2017/03/07452621/13rRUxYIN5A", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqG0SXH", "title": "Computer Graphics and Applications, Pacific Conference on", "acronym": "pg", "groupId": "1000130", "volume": "0", "displayVolume": "0", "year": "2000", "__typename": "ProceedingType" }, "article": { "id": "12OmNzBOhSI", "doi": "10.1109/PCCGA.2000.883864", "title": "Interactive Rendering Method for Displaying Shafts of Light", "normalizedTitle": "Interactive Rendering Method for Displaying Shafts of Light", "abstract": "Recently, graphics hardware has increased in capability, and is now available even on standard PCs. These advances have encouraged researchers to develop hardware-accelerated methods for rendering realistic images. One of the important elements in enhancing reality is the effect of atmospheric scattering. The scattering of light due to atmospheric particles has to be taken into account in order to display shafts of light produced by studio spotlights and headlights of automobiles, for example. The purpose of this paper is to develop a method for displaying shafts of light at interactive rates by making use of the graphics hardware. The method makes use of hardware-accelerated volume rendering techniques to display the shafts of light.", "abstracts": [ { "abstractType": "Regular", "content": "Recently, graphics hardware has increased in capability, and is now available even on standard PCs. These advances have encouraged researchers to develop hardware-accelerated methods for rendering realistic images. One of the important elements in enhancing reality is the effect of atmospheric scattering. The scattering of light due to atmospheric particles has to be taken into account in order to display shafts of light produced by studio spotlights and headlights of automobiles, for example. The purpose of this paper is to develop a method for displaying shafts of light at interactive rates by making use of the graphics hardware. The method makes use of hardware-accelerated volume rendering techniques to display the shafts of light.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Recently, graphics hardware has increased in capability, and is now available even on standard PCs. These advances have encouraged researchers to develop hardware-accelerated methods for rendering realistic images. One of the important elements in enhancing reality is the effect of atmospheric scattering. The scattering of light due to atmospheric particles has to be taken into account in order to display shafts of light produced by studio spotlights and headlights of automobiles, for example. The purpose of this paper is to develop a method for displaying shafts of light at interactive rates by making use of the graphics hardware. The method makes use of hardware-accelerated volume rendering techniques to display the shafts of light.", "fno": "08680031", "keywords": [ "Shafts Of Light", "Graphics Hardware", "Atmospheric Scattering", "Realistic Image Synthesis" ], "authors": [ { "affiliation": "Hokkaido University", "fullName": "Yoshinori Dobashi", "givenName": "Yoshinori", "surname": "Dobashi", "__typename": "ArticleAuthorType" }, { "affiliation": "Hokkaido University", "fullName": "Tsuyoshi Yamamoto", "givenName": "Tsuyoshi", "surname": "Yamamoto", "__typename": "ArticleAuthorType" }, { "affiliation": "University of Tokyo", "fullName": "Tomoyuki Nishita", "givenName": "Tomoyuki", "surname": "Nishita", "__typename": "ArticleAuthorType" } ], "idPrefix": "pg", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "2000-10-01T00:00:00", "pubType": "proceedings", "pages": "31", "year": "2000", "issn": null, "isbn": "0-7695-0868-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "08680023", "articleId": "12OmNzh5z2c", "__typename": "AdjacentArticleType" }, "next": { "fno": "08680040", "articleId": "12OmNBUAvZJ", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvEyR7f", "title": "Electronics, Robotics and Automotive Mechanics Conference", "acronym": "cerma", "groupId": "1001305", "volume": "0", "displayVolume": "0", "year": "2010", "__typename": "ProceedingType" }, "article": { "id": "12OmNzhna82", "doi": "10.1109/CERMA.2010.83", "title": "FPGA Implementation of a 16-Channel Lock-In Laser Light Scattering System", "normalizedTitle": "FPGA Implementation of a 16-Channel Lock-In Laser Light Scattering System", "abstract": "The use of optical techniques for the measurement and characterization of physical phenomena is of great interest in the scientific and industrial communities. This work shows the implementation of an integrated system for measuring Multi-angle Laser Light Scattering on a Field Programmable Gate Array. The embedded design delivers a compact and highly competitive system. The system is able to display, in real-time, the light intensity from a photo detector array showing in this way the corresponding pattern distribution of the scattered light at 16 sensing points. Some laser light scattering patterns from metallic samples with different roughness are shown.", "abstracts": [ { "abstractType": "Regular", "content": "The use of optical techniques for the measurement and characterization of physical phenomena is of great interest in the scientific and industrial communities. This work shows the implementation of an integrated system for measuring Multi-angle Laser Light Scattering on a Field Programmable Gate Array. The embedded design delivers a compact and highly competitive system. The system is able to display, in real-time, the light intensity from a photo detector array showing in this way the corresponding pattern distribution of the scattered light at 16 sensing points. Some laser light scattering patterns from metallic samples with different roughness are shown.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The use of optical techniques for the measurement and characterization of physical phenomena is of great interest in the scientific and industrial communities. This work shows the implementation of an integrated system for measuring Multi-angle Laser Light Scattering on a Field Programmable Gate Array. The embedded design delivers a compact and highly competitive system. The system is able to display, in real-time, the light intensity from a photo detector array showing in this way the corresponding pattern distribution of the scattered light at 16 sensing points. Some laser light scattering patterns from metallic samples with different roughness are shown.", "fno": "4204a721", "keywords": [ "FPGA", "Light Scattering", "Lock In Amplifier", "Surface Roughness" ], "authors": [ { "affiliation": null, "fullName": "Arturo Moreno-Báez", "givenName": "Arturo", "surname": "Moreno-Báez", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Gerardo Miramontes-de Léon", "givenName": "Gerardo", "surname": "Miramontes-de Léon", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Claudia Sifuentes-Gallardo", "givenName": "Claudia", "surname": "Sifuentes-Gallardo", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ernesto García-Domínguez", "givenName": "Ernesto", "surname": "García-Domínguez", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Daniel Alaniz-Lumbreras", "givenName": "Daniel", "surname": "Alaniz-Lumbreras", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jorge A. Huerta-Ruelas", "givenName": "Jorge A.", "surname": "Huerta-Ruelas", "__typename": "ArticleAuthorType" } ], "idPrefix": "cerma", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2010-09-01T00:00:00", "pubType": "proceedings", "pages": "721-725", "year": "2010", "issn": null, "isbn": "978-0-7695-4204-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "4204a717", "articleId": "12OmNBCZnQO", "__typename": "AdjacentArticleType" }, "next": { "fno": "4204a726", "articleId": "12OmNyz5JYf", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/isdea/2010/8333/2/05743412", "title": "Research on Measuring the Surface Texture Based on Laser Light Scattering Method", "doi": null, "abstractUrl": "/proceedings-article/isdea/2010/05743412/12OmNAoUT4h", "parentPublication": { "id": "proceedings/isdea/2010/8333/2", "title": "2010 International Conference on Intelligent System Design and Engineering Application", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/bibe/2017/1324/0/132401a255", "title": "A Simulation Study on Light Scattering Effect on Water-borne Bacteriophage Virus Using Mie Analysis", "doi": null, "abstractUrl": "/proceedings-article/bibe/2017/132401a255/12OmNClQ0tT", "parentPublication": { "id": "proceedings/bibe/2017/1324/0", "title": "2017 IEEE 17th International Conference on Bioinformatics and Bioengineering (BIBE)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209e382", "title": "Light Transport Refocusing for Unknown Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209e382/12OmNqzu6Nb", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wgec/2009/3899/0/3899a060", "title": "Study on Detector Arrangement of Single Cell Counter Based on Forward Light Scattering", "doi": null, "abstractUrl": "/proceedings-article/wgec/2009/3899a060/12OmNrIJqyE", "parentPublication": { "id": "proceedings/wgec/2009/3899/0", "title": "Genetic and Evolutionary Computing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sitis/2013/3211/0/3211a093", "title": "Multiple-Scattering Optical Tomography with Layered Material", "doi": null, "abstractUrl": "/proceedings-article/sitis/2013/3211a093/12OmNzRZpYR", "parentPublication": { "id": "proceedings/sitis/2013/3211/0", "title": "2013 International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmtma/2011/4296/1/4296a704", "title": "Characterization of Nanoparticle Based on the Power Spectrum Density of Dynamic Light Scattering", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2011/4296a704/12OmNzmclAD", "parentPublication": { "id": "proceedings/icmtma/2011/4296/1", "title": "2011 Third International Conference on Measuring Technology and Mechatronics Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sc/1990/2056/0/00130087", "title": "Massively parallel computational methods in light scattering by small particles", "doi": null, "abstractUrl": "/proceedings-article/sc/1990/00130087/12OmNzmclKg", "parentPublication": { "id": "proceedings/sc/1990/2056/0", "title": "SC Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/02/v0342", "title": "Light Scattering from Filaments", "doi": null, "abstractUrl": "/journal/tg/2007/02/v0342/13rRUwI5TXt", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/07/08600345", "title": "Precomputed Multiple Scattering for Rapid Light Simulation in Participating Media", "doi": null, "abstractUrl": "/journal/tg/2020/07/08600345/17D45Xh13tH", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09715049", "title": "Collimated Whole Volume Light Scattering in Homogeneous Finite Media", "doi": null, "abstractUrl": "/journal/tg/5555/01/09715049/1B2DbhImWwE", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNwdbUZf", "title": "Computer Vision, IEEE International Conference on", "acronym": "iccv", "groupId": "1000149", "volume": "1", "displayVolume": "1", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "12OmNzvz6Oz", "doi": "10.1109/ICCV.2005.232", "title": "Structured Light in Scattering Media", "normalizedTitle": "Structured Light in Scattering Media", "abstract": "Virtually all structured light methods assume that the scene and the sources are immersed in pure air and that light is neither scattered nor absorbed. Recently, however, structured lighting has found growing application in underwater and aerial imaging, where scattering effects cannot be ignored. In this paper, we present a comprehensive analysis of two representative methods - light stripe range scanning and photometric stereo - in the presence of scattering. For both methods, we derive physical models for the appearances of a surface immersed in a scattering medium. Based on these models, we present results on (a) the condition for object detectability in light striping and (b) the number of sources required for photometric stereo. In both cases, we demonstrate that while traditional methods fail when scattering is significant, our methods accurately recover the scene (depths, normals, albedos) as well as the properties of the medium. These results are in turn used to restore the appearances of scenes as if they were captured in clear air. Although we have focused on light striping and photometric stereo, our approach can also be extended to other methods such as grid coding, gated and active polarization imaging.", "abstracts": [ { "abstractType": "Regular", "content": "Virtually all structured light methods assume that the scene and the sources are immersed in pure air and that light is neither scattered nor absorbed. Recently, however, structured lighting has found growing application in underwater and aerial imaging, where scattering effects cannot be ignored. In this paper, we present a comprehensive analysis of two representative methods - light stripe range scanning and photometric stereo - in the presence of scattering. For both methods, we derive physical models for the appearances of a surface immersed in a scattering medium. Based on these models, we present results on (a) the condition for object detectability in light striping and (b) the number of sources required for photometric stereo. In both cases, we demonstrate that while traditional methods fail when scattering is significant, our methods accurately recover the scene (depths, normals, albedos) as well as the properties of the medium. These results are in turn used to restore the appearances of scenes as if they were captured in clear air. Although we have focused on light striping and photometric stereo, our approach can also be extended to other methods such as grid coding, gated and active polarization imaging.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Virtually all structured light methods assume that the scene and the sources are immersed in pure air and that light is neither scattered nor absorbed. Recently, however, structured lighting has found growing application in underwater and aerial imaging, where scattering effects cannot be ignored. In this paper, we present a comprehensive analysis of two representative methods - light stripe range scanning and photometric stereo - in the presence of scattering. For both methods, we derive physical models for the appearances of a surface immersed in a scattering medium. Based on these models, we present results on (a) the condition for object detectability in light striping and (b) the number of sources required for photometric stereo. In both cases, we demonstrate that while traditional methods fail when scattering is significant, our methods accurately recover the scene (depths, normals, albedos) as well as the properties of the medium. These results are in turn used to restore the appearances of scenes as if they were captured in clear air. Although we have focused on light striping and photometric stereo, our approach can also be extended to other methods such as grid coding, gated and active polarization imaging.", "fno": "23340420", "keywords": [], "authors": [ { "affiliation": "Carnegie Mellon University", "fullName": "Srinivasa G. Narasimhan", "givenName": "Srinivasa G.", "surname": "Narasimhan", "__typename": "ArticleAuthorType" }, { "affiliation": "Columbia University", "fullName": "Shree K. Nayar", "givenName": "Shree K.", "surname": "Nayar", "__typename": "ArticleAuthorType" }, { "affiliation": "Columbia University", "fullName": "Bo Sun", "givenName": "Bo", "surname": "Sun", "__typename": "ArticleAuthorType" }, { "affiliation": "Carnegie Mellon University", "fullName": "Sanjeev J. Koppal", "givenName": "Sanjeev J.", "surname": "Koppal", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2005-10-01T00:00:00", "pubType": "proceedings", "pages": "420-427", "year": "2005", "issn": "1550-5499", "isbn": "0-7695-2334-X", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "23340412", "articleId": "12OmNyvGygD", "__typename": "AdjacentArticleType" }, "next": { "fno": "23340428", "articleId": "12OmNAk5HQ3", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/robot/1992/2720/0/00220125", "title": "Photometric stereo using point light sources", "doi": null, "abstractUrl": "/proceedings-article/robot/1992/00220125/12OmNAkEU4j", "parentPublication": { "id": "proceedings/robot/1992/2720/0", "title": "Proceedings 1992 IEEE International Conference on Robotics and Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2017/0733/0/0733b287", "title": "Generating 5D Light Fields in Scattering Media for Representing 3D Images", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2017/0733b287/12OmNAndigF", "parentPublication": { "id": "proceedings/cvprw/2017/0733/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d415", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d415/12OmNBV9Igo", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2014/5118/0/5118c299", "title": "Scattering Parameters and Surface Normals from Homogeneous Translucent Materials Using Photometric Stereo", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2014/5118c299/12OmNscOUiy", "parentPublication": { "id": "proceedings/cvpr/2014/5118/0", "title": "2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2010/6984/0/05540216", "title": "Analysis of light transport in scattering media", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2010/05540216/12OmNxWLTlm", "parentPublication": { "id": "proceedings/cvpr/2010/6984/0", "title": "2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvaui/2016/5870/0/5870a049", "title": "Shape Reconstruction of Objects in Participating Media by Combining Photometric Stereo and Optical Thickness", "doi": null, "abstractUrl": "/proceedings-article/cvaui/2016/5870a049/12OmNyUWQXe", "parentPublication": { "id": "proceedings/cvaui/2016/5870/0", "title": "2016 ICPR 2nd Workshop on Computer Vision for Analysis of Underwater Imagery (CVAUI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/02/v0342", "title": "Light Scattering from Filaments", "doi": null, "abstractUrl": "/journal/tg/2007/02/v0342/13rRUwI5TXt", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2017/09/07577857", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/journal/tp/2017/09/07577857/13rRUxYIMWv", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/07/08600345", "title": "Precomputed Multiple Scattering for Rapid Light Simulation in Participating Media", "doi": null, "abstractUrl": "/journal/tg/2020/07/08600345/17D45Xh13tH", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09715049", "title": "Collimated Whole Volume Light Scattering in Homogeneous Finite Media", "doi": null, "abstractUrl": "/journal/tg/5555/01/09715049/1B2DbhImWwE", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvAS4s4", "title": "Proceedings 1992 IEEE International Conference on Robotics and Automation", "acronym": "robot", "groupId": "1000639", "volume": "0", "displayVolume": "0", "year": "1992", "__typename": "ProceedingType" }, "article": { "id": "12OmNAkEU4j", "doi": "10.1109/ROBOT.1992.220125", "title": "Photometric stereo using point light sources", "normalizedTitle": "Photometric stereo using point light sources", "abstract": "Practical lamps for undersea illumination are closely approximated by point sources, the light from which diverges and also attenuates due to scattering and absorption by the water. The objective is to determine the error introduced by these departures from the usual lighting assumptions of photometric stereo and to find an approach to solution of the brightness equations. It is demonstrated that an iterative approach to photometric stereo, plus a sparse range map containing as little as one range datum for each object in the scene, is sufficient to permit determination of the surface gradients. Simulation indicated that the method improves accuracy and was robust with respect to measurement error in both sparse range data and attenuation length.<>", "abstracts": [ { "abstractType": "Regular", "content": "Practical lamps for undersea illumination are closely approximated by point sources, the light from which diverges and also attenuates due to scattering and absorption by the water. The objective is to determine the error introduced by these departures from the usual lighting assumptions of photometric stereo and to find an approach to solution of the brightness equations. It is demonstrated that an iterative approach to photometric stereo, plus a sparse range map containing as little as one range datum for each object in the scene, is sufficient to permit determination of the surface gradients. Simulation indicated that the method improves accuracy and was robust with respect to measurement error in both sparse range data and attenuation length.<>", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Practical lamps for undersea illumination are closely approximated by point sources, the light from which diverges and also attenuates due to scattering and absorption by the water. The objective is to determine the error introduced by these departures from the usual lighting assumptions of photometric stereo and to find an approach to solution of the brightness equations. It is demonstrated that an iterative approach to photometric stereo, plus a sparse range map containing as little as one range datum for each object in the scene, is sufficient to permit determination of the surface gradients. Simulation indicated that the method improves accuracy and was robust with respect to measurement error in both sparse range data and attenuation length.", "fno": "00220125", "keywords": [ "Brightness", "Image Processing", "Iterative Methods", "Lighting", "Iterative Method", "Image Processing", "Point Light Sources", "Photometric Stereo", "Brightness", "Surface Gradients", "Sparse Range Data", "Attenuation Length", "Photometry", "Light Sources", "Lamps", "Lighting", "Light Scattering", "Absorption", "Water Resources", "Brightness", "Equations", "Iterative Methods" ], "authors": [ { "affiliation": "Dept. of Mech. Eng., Hawaii Univ., HI, USA", "fullName": "N. Kolagani", "givenName": "N.", "surname": "Kolagani", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "J.S. Fox", "givenName": "J.S.", "surname": "Fox", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "D.R. Blidberg", "givenName": "D.R.", "surname": "Blidberg", "__typename": "ArticleAuthorType" } ], "idPrefix": "robot", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "1992-01-01T00:00:00", "pubType": "proceedings", "pages": "1759,1760,1761,1762,1763,1764", "year": "1992", "issn": null, "isbn": null, "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "00220124", "articleId": "12OmNx76TH5", "__typename": "AdjacentArticleType" }, "next": { "fno": "00220126", "articleId": "12OmNyrqzqk", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/1992/2855/0/00223147", "title": "Shape reconstruction from photometric stereo", "doi": null, "abstractUrl": "/proceedings-article/cvpr/1992/00223147/12OmNAWH9Gg", "parentPublication": { "id": "proceedings/cvpr/1992/2855/0", "title": "Proceedings 1992 IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2017/0457/0/0457e521", "title": "Semi-Calibrated Near Field Photometric Stereo", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457e521/12OmNB0X8uV", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2014/5118/0/5118c259", "title": "Backscatter Compensated Photometric Stereo with 3 Sources", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2014/5118c259/12OmNBPtJCx", "parentPublication": { "id": "proceedings/cvpr/2014/5118/0", "title": "2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2018/2526/0/08368465", "title": "Near-light photometric stereo using circularly placed point light sources", "doi": null, "abstractUrl": "/proceedings-article/iccp/2018/08368465/12OmNqBbHSi", "parentPublication": { "id": "proceedings/iccp/2018/2526/0", "title": "2018 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/robot/1991/2163/0/00131737", "title": "Automatic planning of light source and camera placement for an active photometric stereo system", "doi": null, "abstractUrl": "/proceedings-article/robot/1991/00131737/12OmNqyUUHw", "parentPublication": { "id": "proceedings/robot/1991/2163/0", "title": "Proceedings. 1991 IEEE International Conference on Robotics and Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/1988/0862/0/00196280", "title": "Calculation of surface position and orientation using the photometric stereo method", "doi": null, "abstractUrl": "/proceedings-article/cvpr/1988/00196280/12OmNwNeYCs", "parentPublication": { "id": "proceedings/cvpr/1988/0862/0", "title": "Proceedings CVPR '88: The Computer Society Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2014/7000/1/7000a115", "title": "Close-Range Photometric Stereo with Point Light Sources", "doi": null, "abstractUrl": "/proceedings-article/3dv/2014/7000a115/12OmNx3ZjoX", "parentPublication": { "id": "proceedings/3dv/2014/7000/2", "title": "2014 2nd International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/robot/1989/1938/0/00099964", "title": "Determining surface curvature with photometric stereo", "doi": null, "abstractUrl": "/proceedings-article/robot/1989/00099964/12OmNzdoMln", "parentPublication": { "id": "proceedings/robot/1989/1938/0", "title": "1989 IEEE International Conference on Robotics and Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2020/01/08478369", "title": "Semi-Calibrated Photometric Stereo", "doi": null, "abstractUrl": "/journal/tp/2020/01/08478369/141AnpAbeCh", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2022/0915/0/091500a317", "title": "Symmetric-light Photometric Stereo", "doi": null, "abstractUrl": "/proceedings-article/wacv/2022/091500a317/1B12PputGEg", "parentPublication": { "id": "proceedings/wacv/2022/0915/0", "title": "2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNAQJzKb", "title": "2015 IEEE Pacific Visualization Symposium (PacificVis)", "acronym": "pacificvis", "groupId": "1001657", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNvDZF6A", "doi": "10.1109/PACIFICVIS.2015.7156382", "title": "Efficient volume illumination with multiple light sources through selective light updates", "normalizedTitle": "Efficient volume illumination with multiple light sources through selective light updates", "abstract": "Incorporating volumetric illumination into rendering of volumetric data increases visual realism, which can lead to improved spatial comprehension. It is known that spatial comprehension can be further improved by incorporating multiple light sources. However, many volumetric illumination algorithms have severe drawbacks when dealing with multiple light sources. These drawbacks are mainly high performance penalties and memory usage, which can be tackled with specialized data structures or data under sampling. In contrast, in this paper we present a method which enables volumetric illumination with multiple light sources without requiring precomputation or impacting visual quality. To achieve this goal, we introduce selective light updates which minimize the required computations when light settings are changed. We will discuss and analyze the novel concepts underlying selective light updates, and demonstrate them when applied to real-world data under different light settings.", "abstracts": [ { "abstractType": "Regular", "content": "Incorporating volumetric illumination into rendering of volumetric data increases visual realism, which can lead to improved spatial comprehension. It is known that spatial comprehension can be further improved by incorporating multiple light sources. However, many volumetric illumination algorithms have severe drawbacks when dealing with multiple light sources. These drawbacks are mainly high performance penalties and memory usage, which can be tackled with specialized data structures or data under sampling. In contrast, in this paper we present a method which enables volumetric illumination with multiple light sources without requiring precomputation or impacting visual quality. To achieve this goal, we introduce selective light updates which minimize the required computations when light settings are changed. We will discuss and analyze the novel concepts underlying selective light updates, and demonstrate them when applied to real-world data under different light settings.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Incorporating volumetric illumination into rendering of volumetric data increases visual realism, which can lead to improved spatial comprehension. It is known that spatial comprehension can be further improved by incorporating multiple light sources. However, many volumetric illumination algorithms have severe drawbacks when dealing with multiple light sources. These drawbacks are mainly high performance penalties and memory usage, which can be tackled with specialized data structures or data under sampling. In contrast, in this paper we present a method which enables volumetric illumination with multiple light sources without requiring precomputation or impacting visual quality. To achieve this goal, we introduce selective light updates which minimize the required computations when light settings are changed. We will discuss and analyze the novel concepts underlying selective light updates, and demonstrate them when applied to real-world data under different light settings.", "fno": "07156382", "keywords": [ "Light Sources", "Lighting", "Rendering Computer Graphics", "Visualization", "Solid Modeling", "Photography", "Color", "Texture", "I 3 7 Computer Graphics Three Dimensional Graphics And Realism Color", "Shading", "Shadowing" ], "authors": [ { "affiliation": "Interactive Visualization Group, Linköping University, Sweden", "fullName": "Erik Sunden", "givenName": "Erik", "surname": "Sunden", "__typename": "ArticleAuthorType" }, { "affiliation": "Visual Computing Research Group, Ulm University, Germany", "fullName": "Timo Ropinski", "givenName": "Timo", "surname": "Ropinski", "__typename": "ArticleAuthorType" } ], "idPrefix": "pacificvis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-04-01T00:00:00", "pubType": "proceedings", "pages": "231-238", "year": "2015", "issn": null, "isbn": "978-1-4673-6879-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "07156381", "articleId": "12OmNBtUdJY", "__typename": "AdjacentArticleType" }, "next": { "fno": "07156383", "articleId": "12OmNz2C1sz", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/robot/1992/2720/0/00220125", "title": "Photometric stereo using point light sources", "doi": null, "abstractUrl": "/proceedings-article/robot/1992/00220125/12OmNAkEU4j", "parentPublication": { "id": "proceedings/robot/1992/2720/0", "title": "Proceedings 1992 IEEE International Conference on Robotics and Automation", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/imstw/2015/6732/0/07177862", "title": "Considerations for light sources: For semiconductor light sensor test", "doi": null, "abstractUrl": "/proceedings-article/imstw/2015/07177862/12OmNClQ0qG", "parentPublication": { "id": "proceedings/imstw/2015/6732/0", "title": "2015 20th International Mixed-Signal Testing Workshop (IMSTW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pbmcv/1995/7021/0/00514673", "title": "A ray-based computational model of light sources and illumination", "doi": null, "abstractUrl": "/proceedings-article/pbmcv/1995/00514673/12OmNrAMEYe", "parentPublication": { "id": "proceedings/pbmcv/1995/7021/0", "title": "Proceedings of the Workshop on Physics-Based Modeling in Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2012/2216/0/06460571", "title": "Estimation of multiple light sources from specular highlights", "doi": null, "abstractUrl": "/proceedings-article/icpr/2012/06460571/12OmNwFid14", "parentPublication": { "id": "proceedings/icpr/2012/2216/0", "title": "2012 21st International Conference on Pattern Recognition (ICPR 2012)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar/2013/2869/0/06671772", "title": "Delta Light Propagation Volumes for mixed reality", "doi": null, "abstractUrl": "/proceedings-article/ismar/2013/06671772/12OmNwkhTdN", "parentPublication": { "id": "proceedings/ismar/2013/2869/0", "title": "2013 IEEE International Symposium on Mixed and Augmented Reality (ISMAR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2010/06/mcg2010060029", "title": "Advanced Volume Illumination with Unconstrained Light Source Positioning", "doi": null, "abstractUrl": "/magazine/cg/2010/06/mcg2010060029/13rRUNvPLcm", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2013/12/ttg2013122946", "title": "Lighting Design for Globally Illuminated Volume Rendering", "doi": null, "abstractUrl": "/journal/tg/2013/12/ttg2013122946/13rRUwvBy8U", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cs/2016/02/mcs2016020090", "title": "Ambient Volume Illumination", "doi": null, "abstractUrl": "/magazine/cs/2016/02/mcs2016020090/13rRUxASu7D", "parentPublication": { "id": "mags/cs", "title": "Computing in Science & Engineering", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ithings-greencom-cpscom-smartdata/2017/3066/0/08276761", "title": "Global Illumination of Dynamic 3D Scene Based on Light Transport Path Reusing", "doi": null, "abstractUrl": "/proceedings-article/ithings-greencom-cpscom-smartdata/2017/08276761/17D45WYQJam", "parentPublication": { "id": "proceedings/ithings-greencom-cpscom-smartdata/2017/3066/0", "title": "2017 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/5555/01/09887904", "title": "Fast and Accurate Illumination Estimation Using LDR Panoramic Images for Realistic Rendering", "doi": null, "abstractUrl": "/journal/tg/5555/01/09887904/1GBRnHyZ1bW", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyKJiwQ", "title": "2013 IEEE International Conference on Computational Photography (ICCP)", "acronym": "iccp", "groupId": "1800125", "volume": "0", "displayVolume": "0", "year": "2013", "__typename": "ProceedingType" }, "article": { "id": "12OmNzmclka", "doi": "10.1109/ICCPhot.2013.6528300", "title": "Descattering of transmissive observation using Parallel High-Frequency Illumination", "normalizedTitle": "Descattering of transmissive observation using Parallel High-Frequency Illumination", "abstract": "The inner structures of an object can be measured by capturing transmissive images. However, the recorded images of a translucent object tend to be unclear due to strong scattering of light inside the object. In this paper, we propose a descattering approach based on Parallel High-frequency Illumination. We show in this paper that the original high-frequency illumination method and the various extended techniques can be uniformly defined as a separation of overlapped and non-overlapped light rays. Also, we show that transmissive light rays do not overlap each other by constructing a parallel projection/measurement system for performing both illumination and observation. We have developed a measurement system that consists of a camera and projector with telecentric lenses and have evaluated descattering effects by extracting transmissive light rays.", "abstracts": [ { "abstractType": "Regular", "content": "The inner structures of an object can be measured by capturing transmissive images. However, the recorded images of a translucent object tend to be unclear due to strong scattering of light inside the object. In this paper, we propose a descattering approach based on Parallel High-frequency Illumination. We show in this paper that the original high-frequency illumination method and the various extended techniques can be uniformly defined as a separation of overlapped and non-overlapped light rays. Also, we show that transmissive light rays do not overlap each other by constructing a parallel projection/measurement system for performing both illumination and observation. We have developed a measurement system that consists of a camera and projector with telecentric lenses and have evaluated descattering effects by extracting transmissive light rays.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "The inner structures of an object can be measured by capturing transmissive images. However, the recorded images of a translucent object tend to be unclear due to strong scattering of light inside the object. In this paper, we propose a descattering approach based on Parallel High-frequency Illumination. We show in this paper that the original high-frequency illumination method and the various extended techniques can be uniformly defined as a separation of overlapped and non-overlapped light rays. Also, we show that transmissive light rays do not overlap each other by constructing a parallel projection/measurement system for performing both illumination and observation. We have developed a measurement system that consists of a camera and projector with telecentric lenses and have evaluated descattering effects by extracting transmissive light rays.", "fno": "06528300", "keywords": [ "Cameras", "Scattering", "Lighting", "Lenses", "Light Sources", "Mirrors" ], "authors": [ { "affiliation": "Inst. of Sci. & Ind. Res., Osaka Univ., Ibaraki, Japan", "fullName": "K. Tanaka", "givenName": "K.", "surname": "Tanaka", "__typename": "ArticleAuthorType" }, { "affiliation": "Inst. of Sci. & Ind. Res., Osaka Univ., Ibaraki, Japan", "fullName": "Y. Mukaigawa", "givenName": "Y.", "surname": "Mukaigawa", "__typename": "ArticleAuthorType" }, { "affiliation": "Microsoft Res. Asia, Beijing, China", "fullName": "Y. Matsushita", "givenName": "Y.", "surname": "Matsushita", "__typename": "ArticleAuthorType" }, { "affiliation": "Inst. of Sci. & Ind. Res., Osaka Univ., Ibaraki, Japan", "fullName": "Y. Yagi", "givenName": "Y.", "surname": "Yagi", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccp", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2013-04-01T00:00:00", "pubType": "proceedings", "pages": "1-8", "year": "2013", "issn": null, "isbn": "978-1-4673-6463-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "06528298", "articleId": "12OmNzWOBfJ", "__typename": "AdjacentArticleType" }, "next": { "fno": "06528301", "articleId": "12OmNC8uRmb", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cgiv/2014/5720/0/5720a083", "title": "Illumination Invariant Measuring of Skin Pigmentation", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2014/5720a083/12OmNCd2rHp", "parentPublication": { "id": "proceedings/cgiv/2014/5720/0", "title": "2014 11th International Conference on Computer Graphics, Imaging and Visualization (CGIV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209e382", "title": "Light Transport Refocusing for Unknown Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209e382/12OmNqzu6Nb", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pbmcv/1995/7021/0/00514673", "title": "A ray-based computational model of light sources and illumination", "doi": null, "abstractUrl": "/proceedings-article/pbmcv/1995/00514673/12OmNrAMEYe", "parentPublication": { "id": "proceedings/pbmcv/1995/7021/0", "title": "Proceedings of the Workshop on Physics-Based Modeling in Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pacificvis/2015/6879/0/07156382", "title": "Efficient volume illumination with multiple light sources through selective light updates", "doi": null, "abstractUrl": "/proceedings-article/pacificvis/2015/07156382/12OmNvDZF6A", "parentPublication": { "id": "proceedings/pacificvis/2015/6879/0", "title": "2015 IEEE Pacific Visualization Symposium (PacificVis)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/1995/7042/0/70420720", "title": "Color constancy under varying illumination", "doi": null, "abstractUrl": "/proceedings-article/iccv/1995/70420720/12OmNzlUKQX", "parentPublication": { "id": "proceedings/iccv/1995/7042/0", "title": "Computer Vision, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2009/4420/0/05459333", "title": "Shadow cameras: Reciprocal views from illumination masks", "doi": null, "abstractUrl": "/proceedings-article/iccv/2009/05459333/12OmNzlUKuY", "parentPublication": { "id": "proceedings/iccv/2009/4420/0", "title": "2009 IEEE 12th International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391d595", "title": "Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d595/12OmNzt0INA", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07470264", "title": "Extinction-Optimized Volume Illumination", "doi": null, "abstractUrl": "/journal/tg/2017/07/07470264/13rRUwI5TR3", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2018/6100/0/610000b661", "title": "Privacy-Preserving Indoor Localization via Active Scene Illumination", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2018/610000b661/17D45VsBTZG", "parentPublication": { "id": "proceedings/cvprw/2018/6100/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2019/1377/0/08798245", "title": "Shadowless Projector: Suppressing Shadows in Projection Mapping with Micro Mirror Array Plate", "doi": null, "abstractUrl": "/proceedings-article/vr/2019/08798245/1cI6ar8DdyE", "parentPublication": { "id": "proceedings/vr/2019/1377/0", "title": "2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBDyAaZ", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNzt0INA", "doi": "10.1109/ICCV.2015.410", "title": "Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography", "normalizedTitle": "Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography", "abstract": "Time of flight (ToF) cameras use a temporally modulated light source and measure correlation between the reflected light and a sensor modulation pattern, in order to infer scene depth. In this paper, we show that such correlational sensors can also be used to selectively accept or reject light rays from certain scene depths. The basic idea is to carefully select illumination and sensor modulation patterns such that the correlation is non-zero only in the selected depth range -- thus light reflected from objects outside this depth range do not affect the correlational measurements. We demonstrate a prototype depth-selective camera and highlight two potential applications: imaging through scattering media and virtual blue screening. This depthselectivity can be used to reject back-scattering and reflection from media in front of the subjects of interest, thereby significantly enhancing the ability to image through scattering media-critical for applications such as car navigation in fog and rain. Similarly, such depth selectivity can also be utilized as a virtual blue-screen in cinematography by rejecting light reflecting from background, while selectively retaining light contributions from the foreground subject.", "abstracts": [ { "abstractType": "Regular", "content": "Time of flight (ToF) cameras use a temporally modulated light source and measure correlation between the reflected light and a sensor modulation pattern, in order to infer scene depth. In this paper, we show that such correlational sensors can also be used to selectively accept or reject light rays from certain scene depths. The basic idea is to carefully select illumination and sensor modulation patterns such that the correlation is non-zero only in the selected depth range -- thus light reflected from objects outside this depth range do not affect the correlational measurements. We demonstrate a prototype depth-selective camera and highlight two potential applications: imaging through scattering media and virtual blue screening. This depthselectivity can be used to reject back-scattering and reflection from media in front of the subjects of interest, thereby significantly enhancing the ability to image through scattering media-critical for applications such as car navigation in fog and rain. Similarly, such depth selectivity can also be utilized as a virtual blue-screen in cinematography by rejecting light reflecting from background, while selectively retaining light contributions from the foreground subject.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Time of flight (ToF) cameras use a temporally modulated light source and measure correlation between the reflected light and a sensor modulation pattern, in order to infer scene depth. In this paper, we show that such correlational sensors can also be used to selectively accept or reject light rays from certain scene depths. The basic idea is to carefully select illumination and sensor modulation patterns such that the correlation is non-zero only in the selected depth range -- thus light reflected from objects outside this depth range do not affect the correlational measurements. We demonstrate a prototype depth-selective camera and highlight two potential applications: imaging through scattering media and virtual blue screening. This depthselectivity can be used to reject back-scattering and reflection from media in front of the subjects of interest, thereby significantly enhancing the ability to image through scattering media-critical for applications such as car navigation in fog and rain. Similarly, such depth selectivity can also be utilized as a virtual blue-screen in cinematography by rejecting light reflecting from background, while selectively retaining light contributions from the foreground subject.", "fno": "8391d595", "keywords": [ "Cameras", "Lighting", "Scattering", "Rain", "Light Sources", "Modulation" ], "authors": [ { "affiliation": null, "fullName": "Ryuichi Tadano", "givenName": "Ryuichi", "surname": "Tadano", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Adithya Kumar Pediredla", "givenName": "Adithya Kumar", "surname": "Pediredla", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Ashok Veeraraghavan", "givenName": "Ashok", "surname": "Veeraraghavan", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-12-01T00:00:00", "pubType": "proceedings", "pages": "3595-3603", "year": "2015", "issn": "2380-7504", "isbn": "978-1-4673-8391-2", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "8391d586", "articleId": "12OmNxbmSDc", "__typename": "AdjacentArticleType" }, "next": { "fno": "8391d604", "articleId": "12OmNwB2dYl", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iccv/2015/8391/0/8391d415", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391d415/12OmNBV9Igo", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209e382", "title": "Light Transport Refocusing for Unknown Scattering Medium", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209e382/12OmNqzu6Nb", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cw/2017/2089/0/2089a056", "title": "Voxel-Based Interactive Rendering of Translucent Materials under Area Lights Using Sparse Samples", "doi": null, "abstractUrl": "/proceedings-article/cw/2017/2089a056/12OmNvDqsQf", "parentPublication": { "id": "proceedings/cw/2017/2089/0", "title": "2017 International Conference on Cyberworlds (CW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2012/1226/0/047O1A02", "title": "Decomposing global light transport using time of flight imaging", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2012/047O1A02/12OmNwCJOTg", "parentPublication": { "id": "proceedings/cvpr/2012/1226/0", "title": "2012 IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pacificvis/2015/6879/0/07156383", "title": "Advanced lighting for unstructured-grid data visualization", "doi": null, "abstractUrl": "/proceedings-article/pacificvis/2015/07156383/12OmNz2C1sz", "parentPublication": { "id": "proceedings/pacificvis/2015/6879/0", "title": "2015 IEEE Pacific Visualization Symposium (PacificVis)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sitis/2013/3211/0/3211a093", "title": "Multiple-Scattering Optical Tomography with Layered Material", "doi": null, "abstractUrl": "/proceedings-article/sitis/2013/3211a093/12OmNzRZpYR", "parentPublication": { "id": "proceedings/sitis/2013/3211/0", "title": "2013 International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccp/2013/6463/0/06528300", "title": "Descattering of transmissive observation using Parallel High-Frequency Illumination", "doi": null, "abstractUrl": "/proceedings-article/iccp/2013/06528300/12OmNzmclka", "parentPublication": { "id": "proceedings/iccp/2013/6463/0", "title": "2013 IEEE International Conference on Computational Photography (ICCP)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07470264", "title": "Extinction-Optimized Volume Illumination", "doi": null, "abstractUrl": "/journal/tg/2017/07/07470264/13rRUwI5TR3", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2007/02/v0342", "title": "Light Scattering from Filaments", "doi": null, "abstractUrl": "/journal/tg/2007/02/v0342/13rRUwI5TXt", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2017/09/07577857", "title": "Photometric Stereo in a Scattering Medium", "doi": null, "abstractUrl": "/journal/tp/2017/09/07577857/13rRUxYIMWv", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1p1gnrYka5y", "title": "2020 International Conference on Culture-oriented Science & Technology (ICCST)", "acronym": "iccst", "groupId": "1838984", "volume": "0", "displayVolume": "0", "year": "2020", "__typename": "ProceedingType" }, "article": { "id": "1p1goMqefzq", "doi": "10.1109/ICCST50977.2020.00113", "title": "Stage Lighting Simulation Based on Epipolar Sampling", "normalizedTitle": "Stage Lighting Simulation Based on Epipolar Sampling", "abstract": "With the development of modern economy, the stage performance art plays a more and more important role in people's life, and lighting plays a key role in stage design. With the support of multimedia technology, stage design has reached a new level and developed more abundant artistic visual effects. This paper uses ray marching algorithm based on epipolar lines to simulate the spot light, which is the most common light effect on the stage. It improves the design efficiency of stage lighting in a new form and reduces the energy and time consumption in lighting arrangement. The results show that the algorithm can simulate the stage lighting effect in real time.", "abstracts": [ { "abstractType": "Regular", "content": "With the development of modern economy, the stage performance art plays a more and more important role in people's life, and lighting plays a key role in stage design. With the support of multimedia technology, stage design has reached a new level and developed more abundant artistic visual effects. This paper uses ray marching algorithm based on epipolar lines to simulate the spot light, which is the most common light effect on the stage. It improves the design efficiency of stage lighting in a new form and reduces the energy and time consumption in lighting arrangement. The results show that the algorithm can simulate the stage lighting effect in real time.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "With the development of modern economy, the stage performance art plays a more and more important role in people's life, and lighting plays a key role in stage design. With the support of multimedia technology, stage design has reached a new level and developed more abundant artistic visual effects. This paper uses ray marching algorithm based on epipolar lines to simulate the spot light, which is the most common light effect on the stage. It improves the design efficiency of stage lighting in a new form and reduces the energy and time consumption in lighting arrangement. The results show that the algorithm can simulate the stage lighting effect in real time.", "fno": "813800a546", "keywords": [ "Art", "Data Visualisation", "Interactive Systems", "Lighting", "Ray Tracing", "Rendering Computer Graphics", "Stage Lighting Simulation", "Modern Economy", "Stage Performance Art", "Lighting Plays", "Stage Design", "Multimedia Technology", "Abundant Artistic Visual Effects", "Ray Marching Algorithm", "Epipolar Lines", "Spot Light", "Common Light Effect", "Stage Lighting Effect", "Lighting Arrangement", "Design Efficiency", "Scattering", "Light Sources", "Lighting", "Interpolation", "Rendering Computer Graphics", "Cameras", "Art", "Virtual Stage", "Lighting Simulation", "Ray Marching", "Epipolar Sampling" ], "authors": [ { "affiliation": "Communication University of China,School of Information and Communication Engineering,Beijing,China", "fullName": "Jiawen Li", "givenName": "Jiawen", "surname": "Li", "__typename": "ArticleAuthorType" }, { "affiliation": "Communication University of China,School of Information and Communication Engineering,Beijing,China", "fullName": "Yinghua Shen", "givenName": "Yinghua", "surname": "Shen", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccst", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2020-10-01T00:00:00", "pubType": "proceedings", "pages": "546-551", "year": "2020", "issn": null, "isbn": "978-1-7281-8138-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "813800a543", "articleId": "1p1gooanqrC", "__typename": "AdjacentArticleType" }, "next": { "fno": "813800a552", "articleId": "1p1gpoXHyww", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cso/2011/4335/0/4335a777", "title": "Research and Realization of New Stage Lighting Control System", "doi": null, "abstractUrl": "/proceedings-article/cso/2011/4335a777/12OmNBSBkb5", "parentPublication": { "id": "proceedings/cso/2011/4335/0", "title": "2011 Fourth International Joint Conference on Computational Sciences and Optimization", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/vr/2015/1727/0/07223335", "title": "Light field projection for lighting reproduction", "doi": null, "abstractUrl": "/proceedings-article/vr/2015/07223335/12OmNs0C9XZ", "parentPublication": { "id": "proceedings/vr/2015/1727/0", "title": "2015 IEEE Virtual Reality (VR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icat/2013/11/0/06728914", "title": "Parallel lighting and reflectance estimation based on inverse rendering", "doi": null, "abstractUrl": "/proceedings-article/icat/2013/06728914/12OmNxwENCV", "parentPublication": { "id": "proceedings/icat/2013/11/0", "title": "2013 23rd International Conference on Artificial Reality and Telexistence (ICAT)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pacificvis/2015/6879/0/07156383", "title": "Advanced lighting for unstructured-grid data visualization", "doi": null, "abstractUrl": "/proceedings-article/pacificvis/2015/07156383/12OmNz2C1sz", "parentPublication": { "id": "proceedings/pacificvis/2015/6879/0", "title": "2015 IEEE Pacific Visualization Symposium (PacificVis)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2017/6067/0/08019481", "title": "Near-surface lighting estimation and reconstruction", "doi": null, "abstractUrl": "/proceedings-article/icme/2017/08019481/12OmNzWx052", "parentPublication": { "id": "proceedings/icme/2017/6067/0", "title": "2017 IEEE International Conference on Multimedia and Expo (ICME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icat/2017/3337/0/08171605", "title": "3D configurator for portrait lighting simulations", "doi": null, "abstractUrl": "/proceedings-article/icat/2017/08171605/12OmNzn38Lf", "parentPublication": { "id": "proceedings/icat/2017/3337/0", "title": "2017 XXVI International Conference on Information, Communication and Automation Technologies (ICAT)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2020/04/08606207", "title": "Hierarchical Bayesian Inverse Lighting of Portraits with a Virtual Light Stage", "doi": null, "abstractUrl": "/journal/tp/2020/04/08606207/17D45W2WyxK", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icctec/2017/5784/0/578400a295", "title": "Study on LED Color Mixing for Stage Lighting Based on Locus Fitting of Blackbody", "doi": null, "abstractUrl": "/proceedings-article/icctec/2017/578400a295/1cks7sBlY8U", "parentPublication": { "id": "proceedings/icctec/2017/5784/0", "title": "2017 International Conference on Computer Technology, Electronics and Communication (ICCTEC)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icicta/2019/4284/0/428400a112", "title": "Research on the Present Situation and Optimum Analysis of Lighting Energy-Saving Technology in Residential Areas", "doi": null, "abstractUrl": "/proceedings-article/icicta/2019/428400a112/1hQqEQByQuI", "parentPublication": { "id": "proceedings/icicta/2019/4284/0", "title": "2019 12th International Conference on Intelligent Computation Technology and Automation (ICICTA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccea/2020/5904/0/09103777", "title": "Lighting Style and Color Temperature to Emotion Response in Architecture Illumination of the Historic Buildings in Dalian", "doi": null, "abstractUrl": "/proceedings-article/iccea/2020/09103777/1kesDXKTmla", "parentPublication": { "id": "proceedings/iccea/2020/5904/0", "title": "2020 International Conference on Computer Engineering and Application (ICCEA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNCmpcNk", "title": "Visualization Conference, IEEE", "acronym": "ieee-vis", "groupId": "1000796", "volume": "0", "displayVolume": "0", "year": "2005", "__typename": "ProceedingType" }, "article": { "id": "12OmNrkT7FS", "doi": "10.1109/VISUAL.2005.1532824", "title": "Reconstructing manifold and non-manifold surfaces from point clouds", "normalizedTitle": "Reconstructing manifold and non-manifold surfaces from point clouds", "abstract": "This paper presents a novel approach for surface reconstruction from point clouds. The proposed technique is general in the sense that it naturally handles both manifold and non-manifold surfaces, providing a consistent way for reconstructing closed surfaces as well as surfaces with boundaries. It is also robust in the presence of noise, irregular sampling and surface gaps. Furthermore, it is fast, parallelizable and easy to implement because it is based on simple local operations. In this approach, surface reconstruction consists of three major steps: first, the space containing the point cloud is subdivided, creating a voxel representation. Then, a voxel surface is computed using gap filling and topological thinning operations. Finally, the resulting voxel surface is converted into a polygonal mesh. We demonstrate the effectiveness of our approach by reconstructing polygonal models from range scans of real objects as well as from synthetic data.", "abstracts": [ { "abstractType": "Regular", "content": "This paper presents a novel approach for surface reconstruction from point clouds. The proposed technique is general in the sense that it naturally handles both manifold and non-manifold surfaces, providing a consistent way for reconstructing closed surfaces as well as surfaces with boundaries. It is also robust in the presence of noise, irregular sampling and surface gaps. Furthermore, it is fast, parallelizable and easy to implement because it is based on simple local operations. In this approach, surface reconstruction consists of three major steps: first, the space containing the point cloud is subdivided, creating a voxel representation. Then, a voxel surface is computed using gap filling and topological thinning operations. Finally, the resulting voxel surface is converted into a polygonal mesh. We demonstrate the effectiveness of our approach by reconstructing polygonal models from range scans of real objects as well as from synthetic data.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper presents a novel approach for surface reconstruction from point clouds. The proposed technique is general in the sense that it naturally handles both manifold and non-manifold surfaces, providing a consistent way for reconstructing closed surfaces as well as surfaces with boundaries. It is also robust in the presence of noise, irregular sampling and surface gaps. Furthermore, it is fast, parallelizable and easy to implement because it is based on simple local operations. In this approach, surface reconstruction consists of three major steps: first, the space containing the point cloud is subdivided, creating a voxel representation. Then, a voxel surface is computed using gap filling and topological thinning operations. Finally, the resulting voxel surface is converted into a polygonal mesh. We demonstrate the effectiveness of our approach by reconstructing polygonal models from range scans of real objects as well as from synthetic data.", "fno": "01532824", "keywords": [ "Solid Modelling", "Image Reconstruction", "Image Sampling", "Image Representation", "Image Thinning", "Surface Fitting", "Mesh Generation", "Computational Geometry", "Manifold Surface Reconstruction", "Nonmanifold Surface Reconstruction", "Point Cloud", "Voxel Surface Representation", "Gap Filling", "Topological Thinning", "Polygonal Mesh", "Polygonal Model Reconstruction", "Real Object", "Surface Reconstruction", "Clouds", "Noise Robustness", "Sampling Methods", "Computer Graphics", "Topology", "Computational Geometry", "Chromium", "Virtual Reality", "Application Software" ], "authors": [ { "affiliation": "Stony Brook Univ., NY, USA", "fullName": "J. Wang", "givenName": "J.", "surname": "Wang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "M.M. Oliveira", "givenName": "M.M.", "surname": "Oliveira", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "A.E. Kaufman", "givenName": "A.E.", "surname": "Kaufman", "__typename": "ArticleAuthorType" } ], "idPrefix": "ieee-vis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2005-01-01T00:00:00", "pubType": "proceedings", "pages": "415,416,417,418,419,420,421,422", "year": "2005", "issn": null, "isbn": null, "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "27660038", "articleId": "12OmNxb5hu0", "__typename": "AdjacentArticleType" }, "next": { "fno": "27660039", "articleId": "12OmNAoUTua", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/smi/2004/2075/0/20750243", "title": "Approximating Bounded, Non-Orientable Surfaces from Points", "doi": null, "abstractUrl": "/proceedings-article/smi/2004/20750243/12OmNAndigx", "parentPublication": { "id": "proceedings/smi/2004/2075/0", "title": "Proceedings. Shape Modeling International 2004", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2017/1032/0/1032c372", "title": "PolyFit: Polygonal Surface Reconstruction from Point Clouds", "doi": null, "abstractUrl": "/proceedings-article/iccv/2017/1032c372/12OmNBRKwBF", "parentPublication": { "id": "proceedings/iccv/2017/1032/0", "title": "2017 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2004/2158/2/01315226", "title": "Reconstructing open surfaces from unorganized data points", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2004/01315226/12OmNBpVQ2L", "parentPublication": { "id": "proceedings/cvpr/2004/2158/2", "title": "Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004.", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/27660053", "title": "Reconstructing Manifold and Non-Manifold Surfaces from Point Clouds", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/27660053/12OmNxbmSzt", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/1993/3870/0/00378184", "title": "A binocular stereo algorithm for reconstructing sloping, creased, and broken surfaces in the presence of half-occlusion", "doi": null, "abstractUrl": "/proceedings-article/iccv/1993/00378184/12OmNy5hRff", "parentPublication": { "id": "proceedings/iccv/1993/3870/0", "title": "1993 (4th) International Conference on Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/1998/9176/0/91760383", "title": "Converting Sets of Polygons to Manifold Surfaces by Cutting and Stitching", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/1998/91760383/12OmNzUPpvx", "parentPublication": { "id": "proceedings/ieee-vis/1998/9176/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2001/02/v0136", "title": "Cutting and Stitching: Converting Sets of Polygons to Manifold Surfaces", "doi": null, "abstractUrl": "/journal/tg/2001/02/v0136/13rRUwI5UfR", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2013/02/ttg2013020306", "title": "Reconstructing Open Surfaces via Graph-Cuts", "doi": null, "abstractUrl": "/journal/tg/2013/02/ttg2013020306/13rRUy0qnGk", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600g305", "title": "Reconstructing Surfaces for Sparse Point Clouds with On-Surface Priors", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600g305/1H1jpDpUMPS", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ismar-adjunct/2022/5365/0/536500a216", "title": "Automated Reconstruction of 3D Open Surfaces from Sparse Point Clouds", "doi": null, "abstractUrl": "/proceedings-article/ismar-adjunct/2022/536500a216/1J7WhkwWdAA", "parentPublication": { "id": "proceedings/ismar-adjunct/2022/5365/0", "title": "2022 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNBKW9zi", "title": "2006 19th Brazilian Symposium on Computer Graphics and Image Processing", "acronym": "sibgrapi", "groupId": "1000131", "volume": "0", "displayVolume": "0", "year": "2006", "__typename": "ProceedingType" }, "article": { "id": "12OmNzBOi0T", "doi": "10.1109/SIBGRAPI.2006.6", "title": "Adapted Dynamic Meshes for Deformable Surfaces", "normalizedTitle": "Adapted Dynamic Meshes for Deformable Surfaces", "abstract": "Deformable objects play an important role in many applications, such as animation and simulation. Effective computation with deformable surfaces can be achieved through the use of dynamic meshes. In this paper, we introduce a framework for constructing and maintaining a timevarying adapted mesh structure that conforms to the underlying deformable surface. The adaptation function employs error metrics based on stochastic sampling. Our scheme combines normal and tangential geometric correction with refinement and simplification resolution control. Furthermore, it applies to both parametric and implicit surface descriptions. As the result, we obtain a simple and efficient general scheme that can be used for a wide range of computations.", "abstracts": [ { "abstractType": "Regular", "content": "Deformable objects play an important role in many applications, such as animation and simulation. Effective computation with deformable surfaces can be achieved through the use of dynamic meshes. In this paper, we introduce a framework for constructing and maintaining a timevarying adapted mesh structure that conforms to the underlying deformable surface. The adaptation function employs error metrics based on stochastic sampling. Our scheme combines normal and tangential geometric correction with refinement and simplification resolution control. Furthermore, it applies to both parametric and implicit surface descriptions. As the result, we obtain a simple and efficient general scheme that can be used for a wide range of computations.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Deformable objects play an important role in many applications, such as animation and simulation. Effective computation with deformable surfaces can be achieved through the use of dynamic meshes. In this paper, we introduce a framework for constructing and maintaining a timevarying adapted mesh structure that conforms to the underlying deformable surface. The adaptation function employs error metrics based on stochastic sampling. Our scheme combines normal and tangential geometric correction with refinement and simplification resolution control. Furthermore, it applies to both parametric and implicit surface descriptions. As the result, we obtain a simple and efficient general scheme that can be used for a wide range of computations.", "fno": "26860213", "keywords": [], "authors": [ { "affiliation": "LIV . IC . UNICAMP, Campinas, SP . Brazil", "fullName": "Fernando de Goes", "givenName": "Fernando", "surname": "de Goes", "__typename": "ArticleAuthorType" }, { "affiliation": "LIV . IC . UNICAMP, Campinas, SP . Brazil", "fullName": "Felipe P. G. Bergo", "givenName": "Felipe P. G.", "surname": "Bergo", "__typename": "ArticleAuthorType" }, { "affiliation": "LIV . IC . UNICAMP, Campinas, SP . Brazil", "fullName": "Alexandre X. Falcao", "givenName": "Alexandre X.", "surname": "Falcao", "__typename": "ArticleAuthorType" }, { "affiliation": "LIV . IC . UNICAMP, Campinas, SP . Brazil", "fullName": "Siome Goldenstein", "givenName": "Siome", "surname": "Goldenstein", "__typename": "ArticleAuthorType" } ], "idPrefix": "sibgrapi", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2006-10-01T00:00:00", "pubType": "proceedings", "pages": "213-220", "year": "2006", "issn": "1530-1834", "isbn": "0-7695-2686-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "26860205", "articleId": "12OmNxb5hwt", "__typename": "AdjacentArticleType" }, "next": { "fno": "26860221", "articleId": "12OmNx5Yvfs", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/iciap/2001/1183/0/11830459", "title": "Global Optimization of Deformable Surface Meshes Based on Genetic Algorithms", "doi": null, "abstractUrl": "/proceedings-article/iciap/2001/11830459/12OmNApu5Jh", "parentPublication": { "id": "proceedings/iciap/2001/1183/0", "title": "Proceedings ICIAP 2001. 11th International Conference on Image Analysis and Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dagstuhl/1997/0503/0/05030267", "title": "Techniques and Applications of Deformable Surfaces", "doi": null, "abstractUrl": "/proceedings-article/dagstuhl/1997/05030267/12OmNBA9oyB", "parentPublication": { "id": "proceedings/dagstuhl/1997/0503/0", "title": "Dagstuhl '97 - Scientific Visualization Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/1996/7258/0/72580638", "title": "Bayesian face recognition using deformable intensity surfaces", "doi": null, "abstractUrl": "/proceedings-article/cvpr/1996/72580638/12OmNBKmXjl", "parentPublication": { "id": "proceedings/cvpr/1996/7258/0", "title": "Proceedings CVPR IEEE Computer Society Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2007/1630/0/04409125", "title": "Interacting with Projected Media on Deformable Surfaces", "doi": null, "abstractUrl": "/proceedings-article/iccv/2007/04409125/12OmNrkBwnr", "parentPublication": { "id": "proceedings/iccv/2007/1630/0", "title": "2007 11th IEEE International Conference on Computer Vision", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pg/1999/0293/0/02930208", "title": "A Multiscale Deformable Model for Extracting Complex Surfaces from Volume Images", "doi": null, "abstractUrl": "/proceedings-article/pg/1999/02930208/12OmNvkpl6T", "parentPublication": { "id": "proceedings/pg/1999/0293/0", "title": "Computer Graphics and Applications, Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ca/1994/6240/0/00324009", "title": "Image morphing using deformable surfaces", "doi": null, "abstractUrl": "/proceedings-article/ca/1994/00324009/12OmNvoWUYx", "parentPublication": { "id": "proceedings/ca/1994/6240/0", "title": "Proceedings of Computer Animation '94", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscv/1995/7190/0/71900223", "title": "Matching and recognition using deformable intensity surfaces", "doi": null, "abstractUrl": "/proceedings-article/iscv/1995/71900223/12OmNyFU74f", "parentPublication": { "id": "proceedings/iscv/1995/7190/0", "title": "Computer Vision, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2005/03/v0329", "title": "Dynamic Interaction between Deformable Surfaces and Nonsmooth Objects", "doi": null, "abstractUrl": "/journal/tg/2005/03/v0329/13rRUwIF6dE", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/dagstuhl/1997/0503/0/01423122", "title": "Techniques and Applications of Deformable Surfaces", "doi": null, "abstractUrl": "/proceedings-article/dagstuhl/1997/01423122/1h0N42oG7qE", "parentPublication": { "id": "proceedings/dagstuhl/1997/0503/0", "title": "Dagstuhl '97 - Scientific Visualization Conference", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1J2XJb8ZJ9C", "title": "2022 Topological Data Analysis and Visualization (TopoInVis)", "acronym": "topoinvis", "groupId": "1848466", "volume": "0", "displayVolume": "0", "year": "2022", "__typename": "ProceedingType" }, "article": { "id": "1J2XKqOgZI4", "doi": "10.1109/TopoInVis57755.2022.00012", "title": "Jacobi Set Driven Search for Flexible Fiber Surface Extraction", "normalizedTitle": "Jacobi Set Driven Search for Flexible Fiber Surface Extraction", "abstract": "Isosurfaces are an important tool for analysis and visualization of univariate scalar fields. Earlier works have demonstrated the presence of interesting isosurfaces at isovalues close to critical values. This motivated the development of efficient methods for computing individual components of isosurfaces restricted to a region of interest. Generalization of isosurfaces to fiber surfaces and critical points to Jacobi sets has resulted in new approaches for analyzing bivariate scalar fields. Unlike isosurfaces, there exists no output sensitive method for computing fiber surfaces. Existing methods traverse through all the tetrahedra in the domain. In this paper, we propose the use of the Jacobi set to identify fiber surface components of interest and present an output sensitive approach for its computation. The Jacobi edges are used to initiate the search towards seed tetrahedra that contain the fiber surface, thereby reducing the search space. This approach also leads to effective analysis of the bivariate field by supporting the identification of relevant fiber surfaces near Jacobi edges.", "abstracts": [ { "abstractType": "Regular", "content": "Isosurfaces are an important tool for analysis and visualization of univariate scalar fields. Earlier works have demonstrated the presence of interesting isosurfaces at isovalues close to critical values. This motivated the development of efficient methods for computing individual components of isosurfaces restricted to a region of interest. Generalization of isosurfaces to fiber surfaces and critical points to Jacobi sets has resulted in new approaches for analyzing bivariate scalar fields. Unlike isosurfaces, there exists no output sensitive method for computing fiber surfaces. Existing methods traverse through all the tetrahedra in the domain. In this paper, we propose the use of the Jacobi set to identify fiber surface components of interest and present an output sensitive approach for its computation. The Jacobi edges are used to initiate the search towards seed tetrahedra that contain the fiber surface, thereby reducing the search space. This approach also leads to effective analysis of the bivariate field by supporting the identification of relevant fiber surfaces near Jacobi edges.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Isosurfaces are an important tool for analysis and visualization of univariate scalar fields. Earlier works have demonstrated the presence of interesting isosurfaces at isovalues close to critical values. This motivated the development of efficient methods for computing individual components of isosurfaces restricted to a region of interest. Generalization of isosurfaces to fiber surfaces and critical points to Jacobi sets has resulted in new approaches for analyzing bivariate scalar fields. Unlike isosurfaces, there exists no output sensitive method for computing fiber surfaces. Existing methods traverse through all the tetrahedra in the domain. In this paper, we propose the use of the Jacobi set to identify fiber surface components of interest and present an output sensitive approach for its computation. The Jacobi edges are used to initiate the search towards seed tetrahedra that contain the fiber surface, thereby reducing the search space. This approach also leads to effective analysis of the bivariate field by supporting the identification of relevant fiber surfaces near Jacobi edges.", "fno": "935400a049", "keywords": [ "Approximation Theory", "Computational Geometry", "Data Visualisation", "Bivariate Field", "Bivariate Scalar Fields", "Critical Points", "Flexible Fiber Surface Extraction", "Interesting Isosurfaces", "Jacobi Edges", "Jacobi Sets", "Output Sensitive Approach", "Output Sensitive Method", "Relevant Fiber Surfaces", "Univariate Scalar Fields", "Visualization", "Jacobian Matrices", "Human Computer Interaction", "Data Analysis", "Data Mining", "Isosurfaces", "Human Centered Computing", "Visualization", "Visualization Techniques", "Human Centered Computing", "Visualization Application Domains", "Scientific Visualization" ], "authors": [ { "affiliation": "Indian Institute of Science,Bangalore", "fullName": "Mohit Sharma", "givenName": "Mohit", "surname": "Sharma", "__typename": "ArticleAuthorType" }, { "affiliation": "Indian Institute of Science,Bangalore", "fullName": "Vijay Natarajan", "givenName": "Vijay", "surname": "Natarajan", "__typename": "ArticleAuthorType" } ], "idPrefix": "topoinvis", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2022-10-01T00:00:00", "pubType": "proceedings", "pages": "49-58", "year": "2022", "issn": null, "isbn": "978-1-6654-9354-3", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "935400a039", "articleId": "1J2XLG1rYo8", "__typename": "AdjacentArticleType" }, "next": { "fno": "935400a059", "articleId": "1J2XMGAA9XO", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/bia/1994/5802/0/00315851", "title": "The wrapper algorithm: surface extraction and simplification", "doi": null, "abstractUrl": "/proceedings-article/bia/1994/00315851/12OmNAlNiPa", "parentPublication": { "id": "proceedings/bia/1994/5802/0", "title": "Proceedings of IEEE Workshop on Biomedical Image Analysis", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ipdps/2018/4368/0/436801a940", "title": "Convergence Models and Surprising Results for the Asynchronous Jacobi Method", "doi": null, "abstractUrl": "/proceedings-article/ipdps/2018/436801a940/12OmNvSbBLN", "parentPublication": { "id": "proceedings/ipdps/2018/4368/0", "title": "2018 IEEE International Parallel and Distributed Processing Symposium (IPDPS)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccis/2012/4789/0/4789a631", "title": "A General Jacobi Elliptic Function Rational Expansion Method and Its Applications in Nonlinear Wave Equations", "doi": null, "abstractUrl": "/proceedings-article/iccis/2012/4789a631/12OmNyUnEJC", "parentPublication": { "id": "proceedings/iccis/2012/4789/0", "title": "2012 Fourth International Conference on Computational and Information Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cgiv/2017/0852/0/0852a044", "title": "Three Dimensional Face Surface Recognition by Geodesic Distance Using Jacobi Iterations", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2017/0852a044/12OmNz6iOFW", "parentPublication": { "id": "proceedings/cgiv/2017/0852/0", "title": "2017 14th International Conference on Computer Graphics, Imaging and Visualization (CGiV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2005/2766/0/01532825", "title": "Marching diamonds for unstructured meshes", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2005/01532825/12OmNzYeAMV", "parentPublication": { "id": "proceedings/ieee-vis/2005/2766/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/01/07539653", "title": "Direct Multifield Volume Ray Casting of Fiber Surfaces", "doi": null, "abstractUrl": "/journal/tg/2017/01/07539653/13rRUB6Sq0C", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07471499", "title": "Fast and Exact Fiber Surfaces for Tetrahedral Meshes", "doi": null, "abstractUrl": "/journal/tg/2017/07/07471499/13rRUNvgz9V", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/01/07539583", "title": "Jacobi Fiber Surfaces for Bivariate Reeb Space Computation", "doi": null, "abstractUrl": "/journal/tg/2017/01/07539583/13rRUx0xPif", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/topoinvis/2022/9354/0/935400a039", "title": "Reduced Connectivity for Local Bilinear Jacobi Sets", "doi": null, "abstractUrl": "/proceedings-article/topoinvis/2022/935400a039/1J2XLG1rYo8", "parentPublication": { "id": "proceedings/topoinvis/2022/9354/0", "title": "2022 Topological Data Analysis and Visualization (TopoInVis)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/td/2021/06/09325945", "title": "A Parallel Jacobi-Embedded Gauss-Seidel Method", "doi": null, "abstractUrl": "/journal/td/2021/06/09325945/1qpv9HIrwGc", "parentPublication": { "id": "trans/td", "title": "IEEE Transactions on Parallel & Distributed Systems", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNqJ8taC", "title": "Computer Graphics and Applications, Pacific Conference on", "acronym": "pg", "groupId": "1000130", "volume": "0", "displayVolume": "0", "year": "1997", "__typename": "ProceedingType" }, "article": { "id": "12OmNAsTgQO", "doi": "10.1109/PCCGA.1997.626177", "title": "Modification of n-sided patches based on variation of blending functions", "normalizedTitle": "Modification of n-sided patches based on variation of blending functions", "abstract": "By introducing the edge-based blending function, derived from the surface-based function, we can capture the effects of the edges on the shape of the n-sided patch more directly. Designers can manipulate the edge-based blending function, as well as the geometry of the boundary curves, to design and modify the patch shape. We discuss the derivation of the edge-based blending function, the ease of design and the connection of several patches with G/sup 1/ continuity. We demonstrate its validity by modifying rectangular patches according to its changes.", "abstracts": [ { "abstractType": "Regular", "content": "By introducing the edge-based blending function, derived from the surface-based function, we can capture the effects of the edges on the shape of the n-sided patch more directly. Designers can manipulate the edge-based blending function, as well as the geometry of the boundary curves, to design and modify the patch shape. We discuss the derivation of the edge-based blending function, the ease of design and the connection of several patches with G/sup 1/ continuity. We demonstrate its validity by modifying rectangular patches according to its changes.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "By introducing the edge-based blending function, derived from the surface-based function, we can capture the effects of the edges on the shape of the n-sided patch more directly. Designers can manipulate the edge-based blending function, as well as the geometry of the boundary curves, to design and modify the patch shape. We discuss the derivation of the edge-based blending function, the ease of design and the connection of several patches with G/sup 1/ continuity. We demonstrate its validity by modifying rectangular patches according to its changes.", "fno": "80280091", "keywords": [ "Computational Geometry N Sided Patch Modification Blending Function Variation Edge Based Blending Function Surface Based Function Edges Geometry Boundary Curves Patch Shape Design Rectangular Patches Interpolation CAD Surface Fitting" ], "authors": [ { "affiliation": "Comput. Archit. Lab., Aizu Univ., Fukushima, Japan", "fullName": "M. Adachi", "givenName": "M.", "surname": "Adachi", "__typename": "ArticleAuthorType" }, { "affiliation": "Comput. Archit. Lab., Aizu Univ., Fukushima, Japan", "fullName": "K.T. Miura", "givenName": "K.T.", "surname": "Miura", "__typename": "ArticleAuthorType" } ], "idPrefix": "pg", "isOpenAccess": false, "showRecommendedArticles": false, "showBuyMe": true, "hasPdf": true, "pubDate": "1997-10-01T00:00:00", "pubType": "proceedings", "pages": "91", "year": "1997", "issn": null, "isbn": "0-8186-8028-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "80280080", "articleId": "12OmNwqft2z", "__typename": "AdjacentArticleType" }, "next": { "fno": "80280097", "articleId": "12OmNxveNRD", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [], "articleVideos": [] }
{ "proceeding": { "id": "12OmNynsbxl", "title": "2014 2nd International Conference on 3D Vision (3DV)", "acronym": "3dv", "groupId": "1800494", "volume": "1", "displayVolume": "1", "year": "2014", "__typename": "ProceedingType" }, "article": { "id": "12OmNx8Ouv7", "doi": "10.1109/3DV.2014.22", "title": "Merge2-3D: Combining Multiple Normal Maps with 3D Surfaces", "normalizedTitle": "Merge2-3D: Combining Multiple Normal Maps with 3D Surfaces", "abstract": "We propose an approach to enhance rough 3D geometry with fine details obtained from multiple normal maps. We begin with unaligned 2D normal maps and rough geometry, and automatically optimize the alignments through 2-step iterative registration algorithm. We then map the normals onto the surface, correcting and seamlessly blending them together. Finally, we optimize the geometry to produce high-quality 3D models that incorporate the high-frequency details from the normal maps. We demonstrate that our algorithm improves upon the results produced by some well-known algorithms: Poisson surface reconstruction [1] and the algorithm proposed by Nehab et al. [2].", "abstracts": [ { "abstractType": "Regular", "content": "We propose an approach to enhance rough 3D geometry with fine details obtained from multiple normal maps. We begin with unaligned 2D normal maps and rough geometry, and automatically optimize the alignments through 2-step iterative registration algorithm. We then map the normals onto the surface, correcting and seamlessly blending them together. Finally, we optimize the geometry to produce high-quality 3D models that incorporate the high-frequency details from the normal maps. We demonstrate that our algorithm improves upon the results produced by some well-known algorithms: Poisson surface reconstruction [1] and the algorithm proposed by Nehab et al. [2].", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We propose an approach to enhance rough 3D geometry with fine details obtained from multiple normal maps. We begin with unaligned 2D normal maps and rough geometry, and automatically optimize the alignments through 2-step iterative registration algorithm. We then map the normals onto the surface, correcting and seamlessly blending them together. Finally, we optimize the geometry to produce high-quality 3D models that incorporate the high-frequency details from the normal maps. We demonstrate that our algorithm improves upon the results produced by some well-known algorithms: Poisson surface reconstruction [1] and the algorithm proposed by Nehab et al. [2].", "fno": "7000a440", "keywords": [ "Three Dimensional Displays", "Solid Modeling", "Surface Reconstruction", "Geometry", "Rough Surfaces", "Surface Roughness", "Image Reconstruction", "Surface Normals", "2 5 3 D Alignment", "Mesh Enhancement" ], "authors": [ { "affiliation": null, "fullName": "Sema Berkiten", "givenName": "Sema", "surname": "Berkiten", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Xinyi Fan", "givenName": "Xinyi", "surname": "Fan", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Szymon Rusinkiewicz", "givenName": "Szymon", "surname": "Rusinkiewicz", "__typename": "ArticleAuthorType" } ], "idPrefix": "3dv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2014-12-01T00:00:00", "pubType": "proceedings", "pages": "440-447", "year": "2014", "issn": null, "isbn": "978-1-4799-7000-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "7000a432", "articleId": "12OmNyGbI9u", "__typename": "AdjacentArticleType" }, "next": { "fno": "7000a448", "articleId": "12OmNBKEynF", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cgiv/2013/5051/0/5051a027", "title": "Real-Time Rendering of Rough Refraction under Dynamically Varying Environmental Lighting", "doi": null, "abstractUrl": "/proceedings-article/cgiv/2013/5051a027/12OmNAWH9zn", "parentPublication": { "id": "proceedings/cgiv/2013/5051/0", "title": "2013 10th International Conference Computer Graphics, Imaging and Visualization (CGIV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2012/1611/0/06239249", "title": "Realistic 3D reconstruction of the human teeth using shape from shading with shape priors", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2012/06239249/12OmNC0guzp", "parentPublication": { "id": "proceedings/cvprw/2012/1611/0", "title": "2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccis/2012/4789/0/4789a127", "title": "Finite Element Analysis of Normal Contact Stiffness between Real Rough Surfaces Based on ANSYS", "doi": null, "abstractUrl": "/proceedings-article/iccis/2012/4789a127/12OmNCf1Dx3", "parentPublication": { "id": "proceedings/iccis/2012/4789/0", "title": "2012 Fourth International Conference on Computational and Information Sciences", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccvw/2011/0063/0/06130447", "title": "A pixel-based approach to template-based monocular 3D reconstruction of deformable surfaces", "doi": null, "abstractUrl": "/proceedings-article/iccvw/2011/06130447/12OmNCga1RM", "parentPublication": { "id": "proceedings/iccvw/2011/0063/0", "title": "2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icip/1994/6952/2/00413545", "title": "Shape from shading for non-Lambertian surfaces", "doi": null, "abstractUrl": "/proceedings-article/icip/1994/00413545/12OmNvsDHHP", "parentPublication": { "id": "proceedings/icip/1994/6952/2", "title": "Proceedings of 1st International Conference on Image Processing", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icme/2007/1016/0/04284896", "title": "Hole Filling on Three-Dimensional Surface Texture", "doi": null, "abstractUrl": "/proceedings-article/icme/2007/04284896/12OmNy4IF6j", "parentPublication": { "id": "proceedings/icme/2007/1016/0", "title": "2007 International Conference on Multimedia & Expo", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2015/8332/0/8332a019", "title": "Multi-view Reconstruction of Highly Specular Surfaces in Uncontrolled Environments", "doi": null, "abstractUrl": "/proceedings-article/3dv/2015/8332a019/12OmNynJMFo", "parentPublication": { "id": "proceedings/3dv/2015/8332/0", "title": "2015 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2001/1143/1/00937522", "title": "New perspectives on geometric reflection theory from rough surfaces", "doi": null, "abstractUrl": "/proceedings-article/iccv/2001/00937522/12OmNzmclV4", "parentPublication": { "id": "proceedings/iccv/2001/1143/1", "title": "Computer Vision, IEEE International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/1993/3880/0/00341163", "title": "Diffuse reflectance from rough surfaces", "doi": null, "abstractUrl": "/proceedings-article/cvpr/1993/00341163/12OmNzwpU3S", "parentPublication": { "id": "proceedings/cvpr/1993/3880/0", "title": "Proceedings of IEEE Conference on Computer Vision and Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000a283", "title": "GeoNet: Geometric Neural Network for Joint Depth and Surface Normal Estimation", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000a283/17D45VTRonG", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNvlxJwN", "title": "2016 Eighth International Conference on Measuring Technology and Mechatronics Automation (ICMTMA)", "acronym": "icmtma", "groupId": "1002837", "volume": "0", "displayVolume": "0", "year": "2016", "__typename": "ProceedingType" }, "article": { "id": "12OmNyNQSQ4", "doi": "10.1109/ICMTMA.2016.15", "title": "A Point Cloud Model Based Image Relief Effect Design", "normalizedTitle": "A Point Cloud Model Based Image Relief Effect Design", "abstract": "This paper mainly studies the relief extraction and data processing in image fusion design. To depend on plane interception and geometric detail extraction in detail shift of point model, intercepting plane is built and the segmentation between relief and base plane of relief product is realized through plane interception. Then, the extracted intermediate relief is performed the isolated sampling point deletion as well as the re-obtained relief boundary optimization of the cut boundary in relief to finally realize relief extraction of relief product. After relief extraction, the extracted relief is performed data processing, which mainly includes smoothness denoising and simplification. Based on second-order Laplace algorithm, smooth denoising inhibits high frequency component and keeps or even enhances low frequency component and overcomes vertex drift and model tearing through avoiding tangential movement of point.", "abstracts": [ { "abstractType": "Regular", "content": "This paper mainly studies the relief extraction and data processing in image fusion design. To depend on plane interception and geometric detail extraction in detail shift of point model, intercepting plane is built and the segmentation between relief and base plane of relief product is realized through plane interception. Then, the extracted intermediate relief is performed the isolated sampling point deletion as well as the re-obtained relief boundary optimization of the cut boundary in relief to finally realize relief extraction of relief product. After relief extraction, the extracted relief is performed data processing, which mainly includes smoothness denoising and simplification. Based on second-order Laplace algorithm, smooth denoising inhibits high frequency component and keeps or even enhances low frequency component and overcomes vertex drift and model tearing through avoiding tangential movement of point.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper mainly studies the relief extraction and data processing in image fusion design. To depend on plane interception and geometric detail extraction in detail shift of point model, intercepting plane is built and the segmentation between relief and base plane of relief product is realized through plane interception. Then, the extracted intermediate relief is performed the isolated sampling point deletion as well as the re-obtained relief boundary optimization of the cut boundary in relief to finally realize relief extraction of relief product. After relief extraction, the extracted relief is performed data processing, which mainly includes smoothness denoising and simplification. Based on second-order Laplace algorithm, smooth denoising inhibits high frequency component and keeps or even enhances low frequency component and overcomes vertex drift and model tearing through avoiding tangential movement of point.", "fno": "2312a022", "keywords": [ "Three Dimensional Displays", "Solid Modeling", "Mathematical Model", "Surface Treatment", "Noise Reduction", "Data Mining", "Optimization", "Noise", "Point Cloud", "Image Fusion", "Relief", "Extraction" ], "authors": [ { "affiliation": null, "fullName": "Han Minghui", "givenName": "Han", "surname": "Minghui", "__typename": "ArticleAuthorType" } ], "idPrefix": "icmtma", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2016-03-01T00:00:00", "pubType": "proceedings", "pages": "22-25", "year": "2016", "issn": "2157-1481", "isbn": "978-1-5090-2312-7", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "2312a019", "articleId": "12OmNzgNXXA", "__typename": "AdjacentArticleType" }, "next": { "fno": "2312a026", "articleId": "12OmNxwncyj", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvprw/2012/1611/0/06238917", "title": "Similarity based filtering of point clouds", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2012/06238917/12OmNvk7JOA", "parentPublication": { "id": "proceedings/cvprw/2012/1611/0", "title": "2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2013/5099/0/5099a187", "title": "Normal Correction towards Smoothing Point-Based Surfaces", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2013/5099a187/12OmNwFicSu", "parentPublication": { "id": "proceedings/sibgrapi/2013/5099/0", "title": "2013 XXVI Conference on Graphics, Patterns and Images", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2016/5407/0/5407a083", "title": "Robust Feature-Preserving Denoising of 3D Point Clouds", "doi": null, "abstractUrl": "/proceedings-article/3dv/2016/5407a083/12OmNyRxFIQ", "parentPublication": { "id": "proceedings/3dv/2016/5407/0", "title": "2016 Fourth International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icmtma/2015/7143/0/7143a114", "title": "An Improved Relief Display Effect of Image Based on Edge Detection", "doi": null, "abstractUrl": "/proceedings-article/icmtma/2015/7143a114/12OmNzE54Bn", "parentPublication": { "id": "proceedings/icmtma/2015/7143/0", "title": "2015 Seventh International Conference on Measuring Technology and Mechatronics Automation (ICMTMA)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2016/5407/0/5407a118", "title": "Point Cloud Noise and Outlier Removal for Image-Based 3D Reconstruction", "doi": null, "abstractUrl": "/proceedings-article/3dv/2016/5407a118/12OmNzsrwoo", "parentPublication": { "id": "proceedings/3dv/2016/5407/0", "title": "2016 Fourth International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2015/03/06975236", "title": "Bas-Relief Generation and Shape Editing through Gradient-Based Mesh Deformation", "doi": null, "abstractUrl": "/journal/tg/2015/03/06975236/13rRUwI5TXA", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/sibgrapi/2018/9264/0/926400a001", "title": "Adaptive Patches for Mesh Denoising", "doi": null, "abstractUrl": "/proceedings-article/sibgrapi/2018/926400a001/17D45XvMce1", "parentPublication": { "id": "proceedings/sibgrapi/2018/9264/0", "title": "2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icdiime/2022/9009/0/900900a094", "title": "Design and Implementation of Image Relief Based on Computer 3D Modeling", "doi": null, "abstractUrl": "/proceedings-article/icdiime/2022/900900a094/1Iz57hORj8I", "parentPublication": { "id": "proceedings/icdiime/2022/9009/0", "title": "2022 International Conference on 3D Immersion, Interaction and Multi-sensory Experiences (ICDIIME)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/11/08730533", "title": "Multi-Patch Collaborative Point Cloud Denoising via Low-Rank Recovery with Graph Constraint", "doi": null, "abstractUrl": "/journal/tg/2020/11/08730533/1aAxaVT7HtS", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300a052", "title": "Total Denoising: Unsupervised Learning of 3D Point Cloud Cleaning", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300a052/1hVlvo8kleE", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNxWuisc", "title": "2015 28th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)", "acronym": "sibgrapi", "groupId": "1000131", "volume": "0", "displayVolume": "0", "year": "2015", "__typename": "ProceedingType" }, "article": { "id": "12OmNyp9MiX", "doi": "10.1109/SIBGRAPI.2015.45", "title": "Meta-Relief Texture Mapping with Dynamic Texture-Space Ambient Occlusion", "normalizedTitle": "Meta-Relief Texture Mapping with Dynamic Texture-Space Ambient Occlusion", "abstract": "We present an efficient technique for modeling and rendering complex surface details defined at multiple scales. Conceptually, meta-relief texture mapping can be described as recursively mapping finer relief-texture layers on top of coarser ones. Such a factorization has several desirable properties. For instance, it provides a way of simulating highly-complex surface details as a combination of simpler and inexpensive image-based representations. This greatly simplifies the modeling of surface details, enhancing the artists' expressive power. We also introduce a dynamic texture-space ambient-occlusion technique for relief mapping, which greatly improves the quality of relief renderings. We demonstrate the effectiveness of these techniques by creating and rendering a number of meta-relief textures with complex surface details which would have been hard to model directly.", "abstracts": [ { "abstractType": "Regular", "content": "We present an efficient technique for modeling and rendering complex surface details defined at multiple scales. Conceptually, meta-relief texture mapping can be described as recursively mapping finer relief-texture layers on top of coarser ones. Such a factorization has several desirable properties. For instance, it provides a way of simulating highly-complex surface details as a combination of simpler and inexpensive image-based representations. This greatly simplifies the modeling of surface details, enhancing the artists' expressive power. We also introduce a dynamic texture-space ambient-occlusion technique for relief mapping, which greatly improves the quality of relief renderings. We demonstrate the effectiveness of these techniques by creating and rendering a number of meta-relief textures with complex surface details which would have been hard to model directly.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We present an efficient technique for modeling and rendering complex surface details defined at multiple scales. Conceptually, meta-relief texture mapping can be described as recursively mapping finer relief-texture layers on top of coarser ones. Such a factorization has several desirable properties. For instance, it provides a way of simulating highly-complex surface details as a combination of simpler and inexpensive image-based representations. This greatly simplifies the modeling of surface details, enhancing the artists' expressive power. We also introduce a dynamic texture-space ambient-occlusion technique for relief mapping, which greatly improves the quality of relief renderings. We demonstrate the effectiveness of these techniques by creating and rendering a number of meta-relief textures with complex surface details which would have been hard to model directly.", "fno": "7962a001", "keywords": [ "Rendering Computer Graphics", "Fabrics", "Metals", "Surface Texture", "Image Color Analysis", "Geometry", "Solids", "Texture Space Ambient Occlusion", "Meta Relief Mapping", "Multiscale Surface Details", "Real Time Rendering" ], "authors": [ { "affiliation": null, "fullName": "Frederico A. Limberger", "givenName": "Frederico A.", "surname": "Limberger", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Victor C. Schetinger", "givenName": "Victor C.", "surname": "Schetinger", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Manuel M. Oliveira", "givenName": "Manuel M.", "surname": "Oliveira", "__typename": "ArticleAuthorType" } ], "idPrefix": "sibgrapi", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2015-08-01T00:00:00", "pubType": "proceedings", "pages": "1-8", "year": "2015", "issn": "1530-1834", "isbn": "978-1-4673-7962-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "7962z012", "articleId": "12OmNqGA4Yg", "__typename": "AdjacentArticleType" }, "next": { "fno": "7962a009", "articleId": "12OmNBCqbA2", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/aici/2010/4225/3/4225c315", "title": "Approach of Geometric Texture Mapping Based on Discrete Gradient Searching", "doi": null, "abstractUrl": "/proceedings-article/aici/2010/4225c315/12OmNBDgZ0M", "parentPublication": { "id": "proceedings/aici/2010/4225/3", "title": "Artificial Intelligence and Computational Intelligence, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icig/2009/3883/0/3883a547", "title": "Improved Relief Texture Mapping Using Minmax Texture", "doi": null, "abstractUrl": "/proceedings-article/icig/2009/3883a547/12OmNrkjVf8", "parentPublication": { "id": "proceedings/icig/2009/3883/0", "title": "Image and Graphics, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/pg/2002/1784/0/17840156", "title": "Geometric Deformation-Displacement Maps", "doi": null, "abstractUrl": "/proceedings-article/pg/2002/17840156/12OmNwfKjb9", "parentPublication": { "id": "proceedings/pg/2002/1784/0", "title": "Computer Graphics and Applications, Pacific Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iscid/2010/4198/1/4198a007", "title": "Haptic Texture Rendering Using Single Texture Image", "doi": null, "abstractUrl": "/proceedings-article/iscid/2010/4198a007/12OmNxbmSBq", "parentPublication": { "id": "proceedings/iscid/2010/4198/1", "title": "Computational Intelligence and Design, International Symposium on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/ieee-vis/2002/7498/0/7498sheffer", "title": "Seamster: Inconspicuous Low-Distortion Texture Seam Layout", "doi": null, "abstractUrl": "/proceedings-article/ieee-vis/2002/7498sheffer/12OmNylsZUq", "parentPublication": { "id": "proceedings/ieee-vis/2002/7498/0", "title": "Visualization Conference, IEEE", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iciap/1999/0040/0/00401055", "title": "Texture Extraction from Photographs and Rendering with Dynamic Texture Mapping", "doi": null, "abstractUrl": "/proceedings-article/iciap/1999/00401055/12OmNz61drx", "parentPublication": { "id": "proceedings/iciap/1999/0040/0", "title": "Image Analysis and Processing, International Conference on", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2001/1272/1/127210615", "title": "Multiview Texture Models", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2001/127210615/12OmNzTppzh", "parentPublication": { "id": "proceedings/cvpr/2001/1272/1", "title": "Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/1986/11/mcg1986110056", "title": "Survey of Texture Mapping", "doi": null, "abstractUrl": "/magazine/cg/1986/11/mcg1986110056/13rRUxYINas", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2006/03/i0446", "title": "Relief Texture from Specularities", "doi": null, "abstractUrl": "/journal/tp/2006/03/i0446/13rRUy3xY97", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "mags/cg/2005/04/mcg2005040066", "title": "Geometric Texture Modeling", "doi": null, "abstractUrl": "/magazine/cg/2005/04/mcg2005040066/13rRUyZaxsV", "parentPublication": { "id": "mags/cg", "title": "IEEE Computer Graphics and Applications", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyoiYVr", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNAsBFOK", "doi": "10.1109/CVPR.2017.693", "title": "Shape Completion Using 3D-Encoder-Predictor CNNs and Shape Synthesis", "normalizedTitle": "Shape Completion Using 3D-Encoder-Predictor CNNs and Shape Synthesis", "abstract": "We introduce a data-driven approach to complete partial 3D shapes through a combination of volumetric deep neural networks and 3D shape synthesis. From a partially-scanned input shape, our method first infers a low-resolution - but complete - output. To this end, we introduce a 3D-Encoder-Predictor Network (3D-EPN) which is composed of 3D convolutional layers. The network is trained to predict and fill in missing data, and operates on an implicit surface representation that encodes both known and unknown space. This allows us to predict global structure in unknown areas at high accuracy. We then correlate these intermediary results with 3D geometry from a shape database at test time. In a final pass, we propose a patch-based 3D shape synthesis method that imposes the 3D geometry from these retrieved shapes as constraints on the coarsely-completed mesh. This synthesis process enables us to reconstruct fine-scale detail and generate high-resolution output while respecting the global mesh structure obtained by the 3D-EPN. Although our 3D-EPN outperforms state-of-the-art completion method, the main contribution in our work lies in the combination of a data-driven shape predictor and analytic 3D shape synthesis. In our results, we show extensive evaluations on a newly-introduced shape completion benchmark for both real-world and synthetic data.", "abstracts": [ { "abstractType": "Regular", "content": "We introduce a data-driven approach to complete partial 3D shapes through a combination of volumetric deep neural networks and 3D shape synthesis. From a partially-scanned input shape, our method first infers a low-resolution - but complete - output. To this end, we introduce a 3D-Encoder-Predictor Network (3D-EPN) which is composed of 3D convolutional layers. The network is trained to predict and fill in missing data, and operates on an implicit surface representation that encodes both known and unknown space. This allows us to predict global structure in unknown areas at high accuracy. We then correlate these intermediary results with 3D geometry from a shape database at test time. In a final pass, we propose a patch-based 3D shape synthesis method that imposes the 3D geometry from these retrieved shapes as constraints on the coarsely-completed mesh. This synthesis process enables us to reconstruct fine-scale detail and generate high-resolution output while respecting the global mesh structure obtained by the 3D-EPN. Although our 3D-EPN outperforms state-of-the-art completion method, the main contribution in our work lies in the combination of a data-driven shape predictor and analytic 3D shape synthesis. In our results, we show extensive evaluations on a newly-introduced shape completion benchmark for both real-world and synthetic data.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "We introduce a data-driven approach to complete partial 3D shapes through a combination of volumetric deep neural networks and 3D shape synthesis. From a partially-scanned input shape, our method first infers a low-resolution - but complete - output. To this end, we introduce a 3D-Encoder-Predictor Network (3D-EPN) which is composed of 3D convolutional layers. The network is trained to predict and fill in missing data, and operates on an implicit surface representation that encodes both known and unknown space. This allows us to predict global structure in unknown areas at high accuracy. We then correlate these intermediary results with 3D geometry from a shape database at test time. In a final pass, we propose a patch-based 3D shape synthesis method that imposes the 3D geometry from these retrieved shapes as constraints on the coarsely-completed mesh. This synthesis process enables us to reconstruct fine-scale detail and generate high-resolution output while respecting the global mesh structure obtained by the 3D-EPN. Although our 3D-EPN outperforms state-of-the-art completion method, the main contribution in our work lies in the combination of a data-driven shape predictor and analytic 3D shape synthesis. In our results, we show extensive evaluations on a newly-introduced shape completion benchmark for both real-world and synthetic data.", "fno": "0457g545", "keywords": [ "Image Reconstruction", "Image Representation", "Image Resolution", "Learning Artificial Intelligence", "Mesh Generation", "Neural Nets", "Shape Recognition", "Stereo Image Processing", "Coarsely Completed Mesh", "3 D Encoder Predictor CN Ns", "Volumetric Deep Neural Networks", "3 D Encoder Predictor Network", "3 D Convolutional Layers", "Shape Database", "Data Driven Shape Predictor", "Shape Retrieval", "3 D EPN", "Partial 3 D Shape Completion", "Partially Scanned Input Shape", "Network Training", "Implicit Surface Representation", "3 D Geometry", "Fine Scale Detail Reconstruction", "High Resolution Output Generation", "Global Mesh Structure", "Analytic 3 D Shape Synthesis", "Three Dimensional Displays", "Shape", "Databases", "Solid Modeling", "Training", "Geometry" ], "authors": [ { "affiliation": null, "fullName": "Angela Dai", "givenName": "Angela", "surname": "Dai", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Charles Ruizhongtai Qi", "givenName": "Charles Ruizhongtai", "surname": "Qi", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Matthias Nießner", "givenName": "Matthias", "surname": "Nießner", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-07-01T00:00:00", "pubType": "proceedings", "pages": "6545-6554", "year": "2017", "issn": "1063-6919", "isbn": "978-1-5386-0457-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0457g535", "articleId": "12OmNrFkeSI", "__typename": "AdjacentArticleType" }, "next": { "fno": "0457g555", "articleId": "12OmNvzJGch", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/cvpr/2016/8851/0/8851d309", "title": "Learned Binary Spectral Shape Descriptor for 3D Shape Correspondence", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2016/8851d309/12OmNBp52AI", "parentPublication": { "id": "proceedings/cvpr/2016/8851/0", "title": "2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2014/5209/0/5209a052", "title": "LBO-Shape Densities: Efficient 3D Shape Retrieval Using Wavelet Density Estimation", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209a052/12OmNBtl1GV", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2014/5118/0/5118d678", "title": "Are Cars Just 3D Boxes? Jointly Estimating the 3D Shape of Multiple Objects", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2014/5118d678/12OmNzlUKmH", "parentPublication": { "id": "proceedings/cvpr/2014/5118/0", "title": "2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07451283", "title": "Shape Completion from a Single RGBD Image", "doi": null, "abstractUrl": "/journal/tg/2017/07/07451283/13rRUxC0Sw0", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000i629", "title": "Learning Descriptor Networks for 3D Shape Synthesis and Analysis", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000i629/17D45WYQJ7N", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200q6218", "title": "SurfGen: Adversarial 3D Shape Synthesis with Explicit Surface Discriminators", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200q6218/1BmFCScd6yk", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2020/9360/0/09150589", "title": "Deep Octree-based CNNs with Output-Guided Skip Connections for 3D Shape and Scene Completion", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2020/09150589/1lPHnPOCgAU", "parentPublication": { "id": "proceedings/cvprw/2020/9360/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/05/09294054", "title": "Generative VoxelNet: Learning Energy-Based Models for 3D Shape Synthesis and Analysis", "doi": null, "abstractUrl": "/journal/tp/2022/05/09294054/1pA6E1DhtYY", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a101", "title": "KAPLAN: A 3D Point Descriptor for Shape Completion", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a101/1qyxm8oZNf2", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900b768", "title": "Unsupervised 3D Shape Completion through GAN Inversion", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900b768/1yeKDo4SRoc", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNylborE", "title": "2018 IEEE Winter Conference on Applications of Computer Vision (WACV)", "acronym": "wacv", "groupId": "1000040", "volume": "0", "displayVolume": "0", "year": "2018", "__typename": "ProceedingType" }, "article": { "id": "12OmNyKJiqm", "doi": "10.1109/WACV.2018.00099", "title": "DeformNet: Free-Form Deformation Network for 3D Shape Reconstruction from a Single Image", "normalizedTitle": "DeformNet: Free-Form Deformation Network for 3D Shape Reconstruction from a Single Image", "abstract": "3D reconstruction from a single image is a key problem in multiple applications ranging from robotic manipulation to augmented reality. Prior methods have tackled this problem through generative models which predict 3D reconstructions as voxels or point clouds. However, these methods can be computationally expensive and miss fine details. We introduce a new differentiable layer for 3D data deformation and use it in DEFORMNET to learn a model for 3D reconstruction-through-deformation. DEFORMNET takes an image input, finds a nearest shape template from a database, and deforms the template to match the query image. We evaluate our approach on the ShapeNet dataset and show that - (a) the Free-Form Deformation layer is a powerful new building block for Deep Learning models that manipulate 3D data (b) DEFORMNET uses this FFD layer combined with shape retrieval for smooth and detail-preserving 3D reconstruction of qualitatively plausible point clouds with respect to a single query image (c) compared to other state-of-the-art 3D reconstruction methods, DEFORMNET quantitatively matches or outperforms their benchmarks by significant margins.", "abstracts": [ { "abstractType": "Regular", "content": "3D reconstruction from a single image is a key problem in multiple applications ranging from robotic manipulation to augmented reality. Prior methods have tackled this problem through generative models which predict 3D reconstructions as voxels or point clouds. However, these methods can be computationally expensive and miss fine details. We introduce a new differentiable layer for 3D data deformation and use it in DEFORMNET to learn a model for 3D reconstruction-through-deformation. DEFORMNET takes an image input, finds a nearest shape template from a database, and deforms the template to match the query image. We evaluate our approach on the ShapeNet dataset and show that - (a) the Free-Form Deformation layer is a powerful new building block for Deep Learning models that manipulate 3D data (b) DEFORMNET uses this FFD layer combined with shape retrieval for smooth and detail-preserving 3D reconstruction of qualitatively plausible point clouds with respect to a single query image (c) compared to other state-of-the-art 3D reconstruction methods, DEFORMNET quantitatively matches or outperforms their benchmarks by significant margins.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "3D reconstruction from a single image is a key problem in multiple applications ranging from robotic manipulation to augmented reality. Prior methods have tackled this problem through generative models which predict 3D reconstructions as voxels or point clouds. However, these methods can be computationally expensive and miss fine details. We introduce a new differentiable layer for 3D data deformation and use it in DEFORMNET to learn a model for 3D reconstruction-through-deformation. DEFORMNET takes an image input, finds a nearest shape template from a database, and deforms the template to match the query image. We evaluate our approach on the ShapeNet dataset and show that - (a) the Free-Form Deformation layer is a powerful new building block for Deep Learning models that manipulate 3D data (b) DEFORMNET uses this FFD layer combined with shape retrieval for smooth and detail-preserving 3D reconstruction of qualitatively plausible point clouds with respect to a single query image (c) compared to other state-of-the-art 3D reconstruction methods, DEFORMNET quantitatively matches or outperforms their benchmarks by significant margins.", "fno": "488601a858", "keywords": [ "Augmented Reality", "CAD", "Computer Vision", "Image Reconstruction", "Learning Artificial Intelligence", "Object Recognition", "Deep Learning Models", "Free Form Deformation Layer", "Nearest Shape Template", "Image Input", "Reconstruction Through Deformation", "3 D Data Deformation", "Differentiable Layer", "Generative Models", "Augmented Reality", "Robotic Manipulation", "Multiple Applications", "3 D Shape Reconstruction", "Free Form Deformation Network", "Single Query Image", "Qualitatively Plausible Point Clouds", "Smooth Detail Preserving 3 D Reconstruction", "Shape Retrieval", "FFD Layer", "3 D Data DEFORMNET", "Shape", "Three Dimensional Displays", "Strain", "Image Reconstruction", "Solid Modeling", "Databases", "Neural Networks" ], "authors": [ { "affiliation": null, "fullName": "Andrey Kurenkov", "givenName": "Andrey", "surname": "Kurenkov", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Jingwei Ji", "givenName": "Jingwei", "surname": "Ji", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Animesh Garg", "givenName": "Animesh", "surname": "Garg", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Viraj Mehta", "givenName": "Viraj", "surname": "Mehta", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "JunYoung Gwak", "givenName": "JunYoung", "surname": "Gwak", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Christopher Choy", "givenName": "Christopher", "surname": "Choy", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Silvio Savarese", "givenName": "Silvio", "surname": "Savarese", "__typename": "ArticleAuthorType" } ], "idPrefix": "wacv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2018-03-01T00:00:00", "pubType": "proceedings", "pages": "858-866", "year": "2018", "issn": null, "isbn": "978-1-5386-4886-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "488601a848", "articleId": "12OmNA0vnS6", "__typename": "AdjacentArticleType" }, "next": { "fno": "488601a867", "articleId": "12OmNyS6RIt", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icpr/2014/5209/0/5209c257", "title": "3D Face Reconstruction via Feature Point Depth Estimation and Shape Deformation", "doi": null, "abstractUrl": "/proceedings-article/icpr/2014/5209c257/12OmNvwTGE3", "parentPublication": { "id": "proceedings/icpr/2014/5209/0", "title": "2014 22nd International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2017/2610/0/261001a088", "title": "Compact Model Representation for 3D Reconstruction", "doi": null, "abstractUrl": "/proceedings-article/3dv/2017/261001a088/12OmNzUgcWs", "parentPublication": { "id": "proceedings/3dv/2017/2610/0", "title": "2017 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2018/6100/0/610000b170", "title": "Learning Hierarchical Models for Class-Specific Reconstruction from Natural Data", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2018/610000b170/17D45WZZ7DD", "parentPublication": { "id": "proceedings/cvprw/2018/6100/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600b526", "title": "Topologically-Aware Deformation Fields for Single-View 3D Reconstruction", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600b526/1H0Nt2kN3Ms", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600s8511", "title": "Pop-Out Motion: 3D-Aware Image Deformation via Learning the Shape Laplacian", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600s8511/1H1kkAtUvAY", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2020/01/08809843", "title": "DeepOrganNet: On-the-Fly Reconstruction and Visualization of 3D / 4D Lung Models from Single-View Projections by Deep Deformation Network", "doi": null, "abstractUrl": "/journal/tg/2020/01/08809843/1cHEoqU2cj6", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300i627", "title": "GraphX-Convolution for Point Cloud Deformation in 2D-to-3D Conversion", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300i627/1hVl8Nw6brq", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800a618", "title": "In Perfect Shape: Certifiably Optimal 3D Shape Reconstruction From 2D Landmarks", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800a618/1m3nTCMilMs", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/10/09444875", "title": "Variational Autoencoders for Localized Mesh Deformation Component Analysis", "doi": null, "abstractUrl": "/journal/tp/2022/10/09444875/1u51uvab1eM", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/10/09462521", "title": "View-Aware Geometry-Structure Joint Learning for Single-View 3D Shape Reconstruction", "doi": null, "abstractUrl": "/journal/tp/2022/10/09462521/1uDSvbmzJQc", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "12OmNyoiYVr", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "acronym": "cvpr", "groupId": "1000147", "volume": "0", "displayVolume": "0", "year": "2017", "__typename": "ProceedingType" }, "article": { "id": "12OmNzn38Ky", "doi": "10.1109/CVPR.2017.28", "title": "Semantic Scene Completion from a Single Depth Image", "normalizedTitle": "Semantic Scene Completion from a Single Depth Image", "abstract": "This paper focuses on semantic scene completion, a task for producing a complete 3D voxel representation of volumetric occupancy and semantic labels for a scene from a single-view depth map observation. Previous work has considered scene completion and semantic labeling of depth maps separately. However, we observe that these two problems are tightly intertwined. To leverage the coupled nature of these two tasks, we introduce the semantic scene completion network (SSCNet), an end-to-end 3D convolutional network that takes a single depth image as input and simultaneously outputs occupancy and semantic labels for all voxels in the camera view frustum. Our network uses a dilation-based 3D context module to efficiently expand the receptive field and enable 3D context learning. To train our network, we construct SUNCG - a manually created largescale dataset of synthetic 3D scenes with dense volumetric annotations. Our experiments demonstrate that the joint model outperforms methods addressing each task in isolation and outperforms alternative approaches on the semantic scene completion task. The dataset and code is available at http://sscnet.cs.princeton.edu.", "abstracts": [ { "abstractType": "Regular", "content": "This paper focuses on semantic scene completion, a task for producing a complete 3D voxel representation of volumetric occupancy and semantic labels for a scene from a single-view depth map observation. Previous work has considered scene completion and semantic labeling of depth maps separately. However, we observe that these two problems are tightly intertwined. To leverage the coupled nature of these two tasks, we introduce the semantic scene completion network (SSCNet), an end-to-end 3D convolutional network that takes a single depth image as input and simultaneously outputs occupancy and semantic labels for all voxels in the camera view frustum. Our network uses a dilation-based 3D context module to efficiently expand the receptive field and enable 3D context learning. To train our network, we construct SUNCG - a manually created largescale dataset of synthetic 3D scenes with dense volumetric annotations. Our experiments demonstrate that the joint model outperforms methods addressing each task in isolation and outperforms alternative approaches on the semantic scene completion task. The dataset and code is available at http://sscnet.cs.princeton.edu.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "This paper focuses on semantic scene completion, a task for producing a complete 3D voxel representation of volumetric occupancy and semantic labels for a scene from a single-view depth map observation. Previous work has considered scene completion and semantic labeling of depth maps separately. However, we observe that these two problems are tightly intertwined. To leverage the coupled nature of these two tasks, we introduce the semantic scene completion network (SSCNet), an end-to-end 3D convolutional network that takes a single depth image as input and simultaneously outputs occupancy and semantic labels for all voxels in the camera view frustum. Our network uses a dilation-based 3D context module to efficiently expand the receptive field and enable 3D context learning. To train our network, we construct SUNCG - a manually created largescale dataset of synthetic 3D scenes with dense volumetric annotations. Our experiments demonstrate that the joint model outperforms methods addressing each task in isolation and outperforms alternative approaches on the semantic scene completion task. The dataset and code is available at http://sscnet.cs.princeton.edu.", "fno": "0457a190", "keywords": [ "Image Representation", "Learning Artificial Intelligence", "Single Depth Image", "Complete 3 D Voxel Representation", "Semantic Labels", "Single View Depth Map Observation", "Semantic Scene Completion Network", "End To End 3 D Convolutional Network", "3 D Context Learning", "Volumetric Occupancy", "Dilation Based 3 D Context Module", "Three Dimensional Displays", "Semantics", "Solid Modeling", "Shape", "Convolution", "Geometry" ], "authors": [ { "affiliation": null, "fullName": "Shuran Song", "givenName": "Shuran", "surname": "Song", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Fisher Yu", "givenName": "Fisher", "surname": "Yu", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Andy Zeng", "givenName": "Andy", "surname": "Zeng", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Angel X. Chang", "givenName": "Angel X.", "surname": "Chang", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Manolis Savva", "givenName": "Manolis", "surname": "Savva", "__typename": "ArticleAuthorType" }, { "affiliation": null, "fullName": "Thomas Funkhouser", "givenName": "Thomas", "surname": "Funkhouser", "__typename": "ArticleAuthorType" } ], "idPrefix": "cvpr", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2017-07-01T00:00:00", "pubType": "proceedings", "pages": "190-198", "year": "2017", "issn": "1063-6919", "isbn": "978-1-5386-0457-1", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "0457a180", "articleId": "12OmNznkKbV", "__typename": "AdjacentArticleType" }, "next": { "fno": "0457a199", "articleId": "12OmNBOllq4", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dv/2018/8425/0/842500a426", "title": "Adversarial Semantic Scene Completion from a Single Depth Image", "doi": null, "abstractUrl": "/proceedings-article/3dv/2018/842500a426/17D45W9KVHL", "parentPublication": { "id": "proceedings/3dv/2018/8425/0", "title": "2018 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2018/6420/0/642000e578", "title": "ScanComplete: Large-Scale Scene Completion and Semantic Segmentation for 3D Scans", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2018/642000e578/17D45WZZ7ET", "parentPublication": { "id": "proceedings/cvpr/2018/6420/0", "title": "2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300i607", "title": "ForkNet: Multi-Branch Volumetric Semantic Completion From a Single Depth Image", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300i607/1hQqscNGkTu", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvprw/2019/2506/0/250600a416", "title": "Two Stream 3D Semantic Scene Completion", "doi": null, "abstractUrl": "/proceedings-article/cvprw/2019/250600a416/1iTvpmQNrbO", "parentPublication": { "id": "proceedings/cvprw/2019/2506/0", "title": "2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/7.168E197", "title": "3D Sketch-Aware Semantic Scene Completion via Semi-Supervised Structure Prior", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/7.168E197/1m3ngObnCda", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800d348", "title": "Anisotropic Convolutional Networks for 3D Semantic Scene Completion", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800d348/1m3nnvKHeVi", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/3dv/2020/8128/0/812800a801", "title": "SCFusion: Real-time Incremental Scene Reconstruction with Semantic Completion", "doi": null, "abstractUrl": "/proceedings-article/3dv/2020/812800a801/1qyxiNprAo8", "parentPublication": { "id": "proceedings/3dv/2020/8128/0", "title": "2020 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/2021/8808/0/09413252", "title": "EdgeNet: Semantic Scene Completion from a Single RGB- D Image", "doi": null, "abstractUrl": "/proceedings-article/icpr/2021/09413252/1tmjb13jEDS", "parentPublication": { "id": "proceedings/icpr/2021/8808/0", "title": "2020 25th International Conference on Pattern Recognition (ICPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/10/09477025", "title": "Semantic Scene Completion Using Local Deep Implicit Functions on LiDAR Data", "doi": null, "abstractUrl": "/journal/tp/2022/10/09477025/1v2M4ngAF9K", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2021/4509/0/450900a324", "title": "Semantic Scene Completion via Integrating Instances and Scene in-the-Loop", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2021/450900a324/1yeLH3ZxKaQ", "parentPublication": { "id": "proceedings/cvpr/2021/4509/0", "title": "2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1BmEezmpGrm", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "acronym": "iccv", "groupId": "1000149", "volume": "0", "displayVolume": "0", "year": "2021", "__typename": "ProceedingType" }, "article": { "id": "1BmFAk4MRe8", "doi": "10.1109/ICCV48922.2021.01236", "title": "Patch2CAD: Patchwise Embedding Learning for In-the-Wild Shape Retrieval from a Single Image", "normalizedTitle": "Patch2CAD: Patchwise Embedding Learning for In-the-Wild Shape Retrieval from a Single Image", "abstract": "3D perception of object shapes from RGB image input is fundamental towards semantic scene understanding, grounding image-based perception in our spatially 3dimensional real-world environments. To achieve a mapping between image views of objects and 3D shapes, we leverage CAD model priors from existing large-scale databases, and propose a novel approach towards constructing a joint embedding space between 2D images and 3D CAD models in a patch-wise fashion &#x2013; establishing correspondences between patches of an image view of an object and patches of CAD geometry. This enables part similarity reasoning for retrieving similar CADs to a new image view without exact matches in the database. Our patch embedding provides more robust CAD retrieval for shape estimation in our end-to-end estimation of CAD model shape and pose for detected objects in a single input image. Experiments on in-the-wild, complex imagery from ScanNet show that our approach is more robust than state of the art in real-world scenarios without any exact CAD matches.", "abstracts": [ { "abstractType": "Regular", "content": "3D perception of object shapes from RGB image input is fundamental towards semantic scene understanding, grounding image-based perception in our spatially 3dimensional real-world environments. To achieve a mapping between image views of objects and 3D shapes, we leverage CAD model priors from existing large-scale databases, and propose a novel approach towards constructing a joint embedding space between 2D images and 3D CAD models in a patch-wise fashion &#x2013; establishing correspondences between patches of an image view of an object and patches of CAD geometry. This enables part similarity reasoning for retrieving similar CADs to a new image view without exact matches in the database. Our patch embedding provides more robust CAD retrieval for shape estimation in our end-to-end estimation of CAD model shape and pose for detected objects in a single input image. Experiments on in-the-wild, complex imagery from ScanNet show that our approach is more robust than state of the art in real-world scenarios without any exact CAD matches.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "3D perception of object shapes from RGB image input is fundamental towards semantic scene understanding, grounding image-based perception in our spatially 3dimensional real-world environments. To achieve a mapping between image views of objects and 3D shapes, we leverage CAD model priors from existing large-scale databases, and propose a novel approach towards constructing a joint embedding space between 2D images and 3D CAD models in a patch-wise fashion – establishing correspondences between patches of an image view of an object and patches of CAD geometry. This enables part similarity reasoning for retrieving similar CADs to a new image view without exact matches in the database. Our patch embedding provides more robust CAD retrieval for shape estimation in our end-to-end estimation of CAD model shape and pose for detected objects in a single input image. Experiments on in-the-wild, complex imagery from ScanNet show that our approach is more robust than state of the art in real-world scenarios without any exact CAD matches.", "fno": "281200m2569", "keywords": [ "Geometry", "Solid Modeling", "Computer Vision", "Three Dimensional Displays", "Shape", "Databases", "Grounding", "3 D From A Single Image And Shape From X", "Detection And Localization In 2 D And 3 D", "Scene Analysis And Understanding" ], "authors": [ { "affiliation": "Google Research,Brain Team", "fullName": "Weicheng Kuo", "givenName": "Weicheng", "surname": "Kuo", "__typename": "ArticleAuthorType" }, { "affiliation": "Google Research,Brain Team", "fullName": "Anelia Angelova", "givenName": "Anelia", "surname": "Angelova", "__typename": "ArticleAuthorType" }, { "affiliation": "Google Research,Brain Team", "fullName": "Tsung-Yi Lin", "givenName": "Tsung-Yi", "surname": "Lin", "__typename": "ArticleAuthorType" }, { "affiliation": "Technical University of Munich", "fullName": "Angela Dai", "givenName": "Angela", "surname": "Dai", "__typename": "ArticleAuthorType" } ], "idPrefix": "iccv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2021-10-01T00:00:00", "pubType": "proceedings", "pages": "12569-12579", "year": "2021", "issn": null, "isbn": "978-1-6654-2812-5", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "281200m2558", "articleId": "1BmI1xNvy12", "__typename": "AdjacentArticleType" }, "next": { "fno": "281200m2580", "articleId": "1BmEJ6eMaNq", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/3dv/2017/2610/0/261001a383", "title": "Learning Quadrangulated Patches for 3D Shape Parameterization and Completion", "doi": null, "abstractUrl": "/proceedings-article/3dv/2017/261001a383/12OmNApcuxi", "parentPublication": { "id": "proceedings/3dv/2017/2610/0", "title": "2017 International Conference on 3D Vision (3DV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2017/0457/0/0457g545", "title": "Shape Completion Using 3D-Encoder-Predictor CNNs and Shape Synthesis", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2017/0457g545/12OmNAsBFOK", "parentPublication": { "id": "proceedings/cvpr/2017/0457/0", "title": "2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/icpr/1994/6265/1/00576390", "title": "Estimating 3D motion and shape of multiple objects using Hough transform", "doi": null, "abstractUrl": "/proceedings-article/icpr/1994/00576390/12OmNC943Qx", "parentPublication": { "id": "proceedings/icpr/1994/6265/1", "title": "Proceedings of 12th International Conference on Pattern Recognition", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/wacv/2016/0641/0/07477652", "title": "3D shape retrieval using a single depth image from low-cost sensors", "doi": null, "abstractUrl": "/proceedings-article/wacv/2016/07477652/12OmNs0C9SV", "parentPublication": { "id": "proceedings/wacv/2016/0641/0", "title": "2016 IEEE Winter Conference on Applications of Computer Vision (WACV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2015/8391/0/8391c677", "title": "3D-Assisted Feature Synthesis for Novel Views of an Object", "doi": null, "abstractUrl": "/proceedings-article/iccv/2015/8391c677/12OmNwLfMBD", "parentPublication": { "id": "proceedings/iccv/2015/8391/0", "title": "2015 IEEE International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2016/02/07138642", "title": "Shape and Reflectance Estimation in the Wild", "doi": null, "abstractUrl": "/journal/tp/2016/02/07138642/13rRUxBa57o", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tg/2017/07/07451283", "title": "Shape Completion from a Single RGBD Image", "doi": null, "abstractUrl": "/journal/tg/2017/07/07451283/13rRUxC0Sw0", "parentPublication": { "id": "trans/tg", "title": "IEEE Transactions on Visualization & Computer Graphics", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/6.946E17", "title": "ROCA: Robust CAD Model Retrieval and Alignment from a Single Image", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/6.946E17/1H0KzeXQpPO", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2019/4803/0/480300i748", "title": "Joint Embedding of 3D Scan and CAD Objects", "doi": null, "abstractUrl": "/proceedings-article/iccv/2019/480300i748/1hVlH2EylGg", "parentPublication": { "id": "proceedings/iccv/2019/4803/0", "title": "2019 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2022/10/09462521", "title": "View-Aware Geometry-Structure Joint Learning for Single-View 3D Shape Reconstruction", "doi": null, "abstractUrl": "/journal/tp/2022/10/09462521/1uDSvbmzJQc", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }
{ "proceeding": { "id": "1KxUhhFgzlK", "title": "2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)", "acronym": "wacv", "groupId": "1000040", "volume": "0", "displayVolume": "0", "year": "2023", "__typename": "ProceedingType" }, "article": { "id": "1L6LvQR1bs4", "doi": "10.1109/WACV56688.2023.00084", "title": "SIRA: Relightable Avatars from a Single Image", "normalizedTitle": "SIRA: Relightable Avatars from a Single Image", "abstract": "Recovering the geometry of a human head from a single image, while factorizing the materials and illumination, is a severely ill-posed problem that requires prior information to be solved. Methods based on 3D Morphable Models (3DMM), and their combination with differentiable renderers, have shown promising results. However, the expressiveness of 3DMMs is limited, and they typically yield over-smoothed and identity-agnostic 3D shapes limited to the face region. Highly accurate full head reconstructions have recently been obtained with neural fields that parameterize the geometry using multilayer perceptrons. The versatility of these representations has also proved effective for disentangling geometry, materials and lighting. However, these methods require several tens of input images. In this paper, we introduce SIRA, a method which, from a single image, reconstructs human head avatars with high fidelity geometry and factorized lights and surface materials. Our key ingredients are two data-driven statistical models based on neural fields that resolve the ambiguities of single-view 3D surface reconstruction and appearance factorization. Experiments show that SIRA obtains state of the art results in 3D head reconstruction while at the same time it successfully disentangles the global illumination, and the diffuse and specular albedos. Furthermore, our reconstructions are amenable to physically-based appearance editing and head model relighting.", "abstracts": [ { "abstractType": "Regular", "content": "Recovering the geometry of a human head from a single image, while factorizing the materials and illumination, is a severely ill-posed problem that requires prior information to be solved. Methods based on 3D Morphable Models (3DMM), and their combination with differentiable renderers, have shown promising results. However, the expressiveness of 3DMMs is limited, and they typically yield over-smoothed and identity-agnostic 3D shapes limited to the face region. Highly accurate full head reconstructions have recently been obtained with neural fields that parameterize the geometry using multilayer perceptrons. The versatility of these representations has also proved effective for disentangling geometry, materials and lighting. However, these methods require several tens of input images. In this paper, we introduce SIRA, a method which, from a single image, reconstructs human head avatars with high fidelity geometry and factorized lights and surface materials. Our key ingredients are two data-driven statistical models based on neural fields that resolve the ambiguities of single-view 3D surface reconstruction and appearance factorization. Experiments show that SIRA obtains state of the art results in 3D head reconstruction while at the same time it successfully disentangles the global illumination, and the diffuse and specular albedos. Furthermore, our reconstructions are amenable to physically-based appearance editing and head model relighting.", "__typename": "ArticleAbstractType" } ], "normalizedAbstract": "Recovering the geometry of a human head from a single image, while factorizing the materials and illumination, is a severely ill-posed problem that requires prior information to be solved. Methods based on 3D Morphable Models (3DMM), and their combination with differentiable renderers, have shown promising results. However, the expressiveness of 3DMMs is limited, and they typically yield over-smoothed and identity-agnostic 3D shapes limited to the face region. Highly accurate full head reconstructions have recently been obtained with neural fields that parameterize the geometry using multilayer perceptrons. The versatility of these representations has also proved effective for disentangling geometry, materials and lighting. However, these methods require several tens of input images. In this paper, we introduce SIRA, a method which, from a single image, reconstructs human head avatars with high fidelity geometry and factorized lights and surface materials. Our key ingredients are two data-driven statistical models based on neural fields that resolve the ambiguities of single-view 3D surface reconstruction and appearance factorization. Experiments show that SIRA obtains state of the art results in 3D head reconstruction while at the same time it successfully disentangles the global illumination, and the diffuse and specular albedos. Furthermore, our reconstructions are amenable to physically-based appearance editing and head model relighting.", "fno": "934600a775", "keywords": [ "Avatars", "Face Recognition", "Image Reconstruction", "Multilayer Perceptrons", "Rendering Computer Graphics", "Solid Modelling", "3 D Head Reconstruction", "3 D Morphable Models", "3 DMM", "Accurate Full Head Reconstructions", "Appearance Factorization", "Data Driven Statistical Models", "Head Model Relighting", "High Fidelity Geometry", "Identity Agnostic 3 D Shapes", "Input Images", "Neural Fields", "Reconstructs Human Head Avatars", "Relightable Avatars", "Single Image", "SIRA", "Surface Materials", "Geometry", "Surface Reconstruction", "Solid Modeling", "Head", "Three Dimensional Displays", "Shape", "Avatars", "Algorithms 3 D Computer Vision", "Biometrics", "Face", "Gesture", "Body Pose", "Computational Photography", "Image And Video Synthesis" ], "authors": [ { "affiliation": "Crisalix SA", "fullName": "Pol Caselles", "givenName": "Pol", "surname": "Caselles", "__typename": "ArticleAuthorType" }, { "affiliation": "Crisalix SA", "fullName": "Eduard Ramon", "givenName": "Eduard", "surname": "Ramon", "__typename": "ArticleAuthorType" }, { "affiliation": "Crisalix SA", "fullName": "Jaime Garcia", "givenName": "Jaime", "surname": "Garcia", "__typename": "ArticleAuthorType" }, { "affiliation": "Universitat Politècnica de Catalunya", "fullName": "Xavier Giro-i-Nieto", "givenName": "Xavier", "surname": "Giro-i-Nieto", "__typename": "ArticleAuthorType" }, { "affiliation": "Institut de Robòtica i Informàtica Industrial, CSIC-UPC", "fullName": "Francesc Moreno-Noguer", "givenName": "Francesc", "surname": "Moreno-Noguer", "__typename": "ArticleAuthorType" }, { "affiliation": "Crisalix SA", "fullName": "Gil Triginer", "givenName": "Gil", "surname": "Triginer", "__typename": "ArticleAuthorType" } ], "idPrefix": "wacv", "isOpenAccess": false, "showRecommendedArticles": true, "showBuyMe": true, "hasPdf": true, "pubDate": "2023-01-01T00:00:00", "pubType": "proceedings", "pages": "775-784", "year": "2023", "issn": null, "isbn": "978-1-6654-9346-8", "notes": null, "notesType": null, "__typename": "ArticleType" }, "webExtras": [], "adjacentArticles": { "previous": { "fno": "934600a764", "articleId": "1KxUzyJqeOI", "__typename": "AdjacentArticleType" }, "next": { "fno": "934600a785", "articleId": "1KxVq4DixWg", "__typename": "AdjacentArticleType" }, "__typename": "AdjacentArticlesType" }, "recommendedArticles": [ { "id": "proceedings/icpr/2012/2216/0/06460312", "title": "Stage-based 3D scene reconstruction from single image", "doi": null, "abstractUrl": "/proceedings-article/icpr/2012/06460312/12OmNBSjIYo", "parentPublication": { "id": "proceedings/icpr/2012/2216/0", "title": "2012 21st International Conference on Pattern Recognition (ICPR 2012)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2004/2158/2/01315139", "title": "View independent human body pose estimation from a single perspective image", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2004/01315139/12OmNzaQoJO", "parentPublication": { "id": "proceedings/cvpr/2004/2158/2", "title": "Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004.", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "trans/tp/2015/04/06888493", "title": "Mirror Surface Reconstruction from a Single Image", "doi": null, "abstractUrl": "/journal/tp/2015/04/06888493/13rRUILc8gn", "parentPublication": { "id": "trans/tp", "title": "IEEE Transactions on Pattern Analysis & Machine Intelligence", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/fg/2021/3176/0/09667017", "title": "Deep Parametric Surfaces for 3D Outfit Reconstruction from Single View Image", "doi": null, "abstractUrl": "/proceedings-article/fg/2021/09667017/1A6BEk9xRaU", "parentPublication": { "id": "proceedings/fg/2021/3176/0", "title": "2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/iccv/2021/2812/0/281200f600", "title": "H3D-Net: Few-Shot High-Fidelity 3D Head Reconstruction", "doi": null, "abstractUrl": "/proceedings-article/iccv/2021/281200f600/1BmG0cJclQA", "parentPublication": { "id": "proceedings/iccv/2021/2812/0", "title": "2021 IEEE/CVF International Conference on Computer Vision (ICCV)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600p5883", "title": "High-Fidelity Human Avatars from a Single RGB Camera", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600p5883/1H1hK72b9Je", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600s8632", "title": "Neural Head Avatars from Monocular RGB Videos", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600s8632/1H1htwlAaNa", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2022/6946/0/694600n3535", "title": "I M Avatar: Implicit Morphable Head Avatars from Videos", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2022/694600n3535/1H1j2BWBE2c", "parentPublication": { "id": "proceedings/cvpr/2022/6946/0", "title": "2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800h707", "title": "Deep 3D Portrait From a Single Image", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800h707/1m3nijIcYta", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" }, { "id": "proceedings/cvpr/2020/7168/0/716800a676", "title": "OASIS: A Large-Scale Dataset for Single Image 3D in the Wild", "doi": null, "abstractUrl": "/proceedings-article/cvpr/2020/716800a676/1m3obA8sShG", "parentPublication": { "id": "proceedings/cvpr/2020/7168/0", "title": "2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "__typename": "ParentPublication" }, "__typename": "RecommendedArticleType" } ], "articleVideos": [] }