Добірка наукової літератури з теми "Scene parsing"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся зі списками актуальних статей, книг, дисертацій, тез та інших наукових джерел на тему "Scene parsing".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Статті в журналах з теми "Scene parsing"
Shi, Hengcan, Hongliang Li, Fanman Meng, Qingbo Wu, Linfeng Xu, and King Ngi Ngan. "Hierarchical Parsing Net: Semantic Scene Parsing From Global Scene to Objects." IEEE Transactions on Multimedia 20, no. 10 (October 2018): 2670–82. http://dx.doi.org/10.1109/tmm.2018.2812600.
Повний текст джерелаCe Liu, J. Yuen, and A. Torralba. "Nonparametric Scene Parsing via Label Transfer." IEEE Transactions on Pattern Analysis and Machine Intelligence 33, no. 12 (December 2011): 2368–82. http://dx.doi.org/10.1109/tpami.2011.131.
Повний текст джерелаZhang, Rui, Sheng Tang, Yongdong Zhang, Jintao Li, and Shuicheng Yan. "Perspective-Adaptive Convolutions for Scene Parsing." IEEE Transactions on Pattern Analysis and Machine Intelligence 42, no. 4 (April 1, 2020): 909–24. http://dx.doi.org/10.1109/tpami.2018.2890637.
Повний текст джерелаXuelong Li, Lichao Mou, and Xiaoqiang Lu. "Scene Parsing From an MAP Perspective." IEEE Transactions on Cybernetics 45, no. 9 (September 2015): 1876–86. http://dx.doi.org/10.1109/tcyb.2014.2361489.
Повний текст джерелаZhang, Botao, Tao Hong, Rong Xiong, and Sergey A. Chepinskiy. "A terrain segmentation method based on pyramid scene parsing-mobile network for outdoor robots." International Journal of Advanced Robotic Systems 18, no. 5 (September 1, 2021): 172988142110486. http://dx.doi.org/10.1177/17298814211048633.
Повний текст джерелаChen, Xiaoyu, Chuan Wang, Jun Lu, Lianfa Bai, and Jing Han. "Road-Scene Parsing Based on Attentional Prototype-Matching." Sensors 22, no. 16 (August 17, 2022): 6159. http://dx.doi.org/10.3390/s22166159.
Повний текст джерелаZhang, Pingping, Wei Liu, Yinjie Lei, Hongyu Wang, and Huchuan Lu. "Deep Multiphase Level Set for Scene Parsing." IEEE Transactions on Image Processing 29 (2020): 4556–67. http://dx.doi.org/10.1109/tip.2019.2957915.
Повний текст джерелаBoutell, Matthew R., Jiebo Luo, and Christopher M. Brown. "Scene Parsing Using Region-Based Generative Models." IEEE Transactions on Multimedia 9, no. 1 (January 2007): 136–46. http://dx.doi.org/10.1109/tmm.2006.886372.
Повний текст джерелаHager, Gregory D., and Ben Wegbreit. "Scene parsing using a prior world model." International Journal of Robotics Research 30, no. 12 (June 3, 2011): 1477–507. http://dx.doi.org/10.1177/0278364911399340.
Повний текст джерелаBu, Shuhui, Pengcheng Han, Zhenbao Liu, and Junwei Han. "Scene parsing using inference Embedded Deep Networks." Pattern Recognition 59 (November 2016): 188–98. http://dx.doi.org/10.1016/j.patcog.2016.01.027.
Повний текст джерелаДисертації з теми "Scene parsing"
Zhao, Hang Ph D. Massachusetts Institute of Technology. "Visual and auditory scene parsing." Thesis, Massachusetts Institute of Technology, 2019. https://hdl.handle.net/1721.1/122101.
Повний текст джерелаThesis: Ph. D. in Mechanical Engineering and Computation, Massachusetts Institute of Technology, Department of Mechanical Engineering, 2019
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 121-132).
Scene parsing is a fundamental topic in computer vision and computational audition, where people develop computational approaches to achieve human perceptual system's ability in understanding scenes, e.g. group visual regions of an image into objects and segregate sound components in a noisy environment. This thesis investigates fully-supervised and self-supervised machine learning approaches to parse visual and auditory signals, including images, videos, and audios. Visual scene parsing refers to densely grouping and labeling of image regions into object concepts. First I build the MIT scene parsing benchmark based on a large scale, densely annotated dataset ADE20K. This benchmark, together with the state-of-the-art models we open source, offers a powerful tool for the research community to solve semantic and instance segmentation tasks. Then I investigate the challenge of parsing a large number of object categories in the wild. An open vocabulary scene parsing model which combines a convolutional neural network with a structured knowledge graph is proposed to address the challenge. Auditory scene parsing refers to recognizing and decomposing sound components in complex auditory environments. I propose a general audio-visual self-supervised learning framework that learns from a large amount of unlabeled internet videos. The learning process discovers the natural synchronization of vision and sounds without human annotation. The learned model achieves the capability to localize sound sources in videos and separate them from mixture. Furthermore, I demonstrate that motion cues in videos are tightly associated with sounds, which help in solving sound localization and separation problems.
by Hang Zhao.
Ph. D. in Mechanical Engineering and Computation
Ph.D.inMechanicalEngineeringandComputation Massachusetts Institute of Technology, Department of Mechanical Engineering
Lan, Cyril. "Urban scene parsing via low-rank texture patches." Thesis, Massachusetts Institute of Technology, 2012. http://hdl.handle.net/1721.1/77536.
Повний текст джерелаCataloged from PDF version of thesis.
Includes bibliographical references (p. 52-55).
Automatic 3-D reconstruction of city scenes from ground, aerial, and satellite imagery is a difficult problem that has seen active research for nearly two decades. The problem is difficult because many algorithms require salient areas in the image to be identified and segmented, a task that is typically done by humans. We propose a pipeline that detects these salient areas using low-rank texture patches. Areas in images such as building facades contain low-rank textures, which are an intrinsic property of the scene and invariant to viewpoint. The pipeline uses these low-rank patches to automatically rectify images and detect and segment out the patches with an energy-minimizing graph cut. The output is then further parameterized to provide useful data to existing 3-D reconstruction methods. The pipeline was evaluated on challenging test images from Microsoft Bing Maps oblique aerial photography and produced an 80% recall and precision with superb empirical results.
by Cyril Lan.
M.Eng.
Tung, Frederick. "Towards large-scale nonparametric scene parsing of images and video." Thesis, University of British Columbia, 2017. http://hdl.handle.net/2429/60790.
Повний текст джерелаScience, Faculty of
Computer Science, Department of
Graduate
Shu, Allen. "Use of shot/scene parsing in generating and browsing video databases." Thesis, Massachusetts Institute of Technology, 1995. http://hdl.handle.net/1721.1/36985.
Повний текст джерелаPan, Hong. "Superparsing with Improved Segmentation Boundaries through Nonparametric Context." Thesis, Université d'Ottawa / University of Ottawa, 2015. http://hdl.handle.net/10393/32329.
Повний текст джерелаMunoz, Daniel. "Inference Machines: Parsing Scenes via Iterated Predictions." Research Showcase @ CMU, 2013. http://repository.cmu.edu/dissertations/305.
Повний текст джерелаTaghavi, Namin Sarah. "Scene Parsing using Multiple Modalities." Phd thesis, 2016. http://hdl.handle.net/1885/116781.
Повний текст джерелаWang, Ren, and 王任. "Transferring Weakly-Supervised Convolutional Networks for Scene Parsing." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/29046824010257775924.
Повний текст джерела國立清華大學
資訊工程學系
103
Deep neural networks have become more and more popular in computer vision because of their powerful ability to extract distinctive image features. In deep neural networks, transfer learning plays an important role to avoid overfitting. In this thesis, we present a clustering-based method to combine fully-labeled data with weakly-labeled data for convolutional networks. By transfer learning, these convolutional networks can be viewed as pre-trained models for another target task. Next, we design a framework of convolutional networks for scene parsing to demonstrate our idea. Preliminary experimental results show that it is helpful to use these pre-trained convolutional networks for transfer learning.
Yu, Jie-Kuan, and 余界寬. "A Scene Parsing and Classification Method for Baseball Videos." Thesis, 2006. http://ndltd.ncl.edu.tw/handle/wrt3k3.
Повний текст джерела國立臺北科技大學
資訊工程系所
94
The thesis proposes a scene parsing and classification system for baseball videos. The system automatically parses baseball video and extracts import scenes with image content analysis. Firstly, the system selects several candidate import scenes by field/cloth color ratio and scene change detection. Secondly, the system utilizes image features, e.g. object motion detection, field and cloth color detection, camera motion parameters, key-frame analysis, and motion-map comparison, etc, to analysis each candidate import scenes. Finally, the system classifies scenes according to above-mentioned features and predefined rules. Subsequently, the system will establish indexes of scenes correspond to the rules in baseball video database.
He, Tong. "Efficient Scene Parsing with Imagery and Point Cloud Data." Thesis, 2020. http://hdl.handle.net/2440/129534.
Повний текст джерелаThesis (Ph.D.) -- University of Adelaide, School of Computer Science, 2020
Книги з теми "Scene parsing"
Museum, of Contemporary Art (Los Angeles Calif ). The social scene: The Ralph M. Parsons Foundation photography collection at the Museum of Contemporary Art, Los Angeles. Los Angeles: Museum of Contemporary Art, 2000.
Знайти повний текст джерелаLevinson, Marjorie, and Marjorie Levinson. Parsing the Frost. Oxford University Press, 2018. http://dx.doi.org/10.1093/oso/9780198810315.003.0009.
Повний текст джерелаColeman, A. D., Cornelia H. Butler, Liz Kotz, and Calif.) Museum of Contemporary Art (Los. The Social Scene, The Ralph R. Parsons Foundation, Photography Collection. Museum of Contemporary Art, 2000.
Знайти повний текст джерелаЧастини книг з теми "Scene parsing"
Liu, Ce, Jenny Yuen, and Antonio Torralba. "Nonparametric Scene Parsing via Label Transfer." In Dense Image Correspondences for Computer Vision, 207–36. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-23048-1_10.
Повний текст джерелаZhong, Guangyu, Yi-Hsuan Tsai, and Ming-Hsuan Yang. "Weakly-Supervised Video Scene Co-parsing." In Computer Vision – ACCV 2016, 20–36. Cham: Springer International Publishing, 2017. http://dx.doi.org/10.1007/978-3-319-54181-5_2.
Повний текст джерелаXiao, Tete, Yingcheng Liu, Bolei Zhou, Yuning Jiang, and Jian Sun. "Unified Perceptual Parsing for Scene Understanding." In Computer Vision – ECCV 2018, 432–48. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-030-01228-1_26.
Повний текст джерелаWu, Tianyi, Yu Lu, Yu Zhu, Chuang Zhang, Ming Wu, Zhanyu Ma, and Guodong Guo. "GINet: Graph Interaction Network for Scene Parsing." In Computer Vision – ECCV 2020, 34–51. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-58520-4_3.
Повний текст джерелаLu, Ye, Xian Zhong, Wenxuan Liu, Jingling Yuan, and Bo Ma. "Tree-Structured Channel-Fuse Network for Scene Parsing." In Advances in Intelligent Systems and Computing, 697–709. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-55180-3_53.
Повний текст джерелаTung, Frederick, and James J. Little. "CollageParsing: Nonparametric Scene Parsing by Adaptive Overlapping Windows." In Computer Vision – ECCV 2014, 511–25. Cham: Springer International Publishing, 2014. http://dx.doi.org/10.1007/978-3-319-10599-4_33.
Повний текст джерелаLi, Xiangtai, Ansheng You, Zhen Zhu, Houlong Zhao, Maoke Yang, Kuiyuan Yang, Shaohua Tan, and Yunhai Tong. "Semantic Flow for Fast and Accurate Scene Parsing." In Computer Vision – ECCV 2020, 775–93. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-58452-8_45.
Повний текст джерелаTang, Keke, Zhe Zhao, and Xiaoping Chen. "Joint Visual Phrase Detection to Boost Scene Parsing." In Advances in Visual Computing, 389–99. Cham: Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-27863-6_36.
Повний текст джерелаYu, Hui, Yuecheng Song, Wenyu Ju, and Zhenbao Liu. "Scene Parsing with Deep Features and Spatial Structure Learning." In Lecture Notes in Computer Science, 715–22. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-48896-7_71.
Повний текст джерелаCui, Xiaofei, Hanbing Qu, Xi Chen, Ziliang Qi, and Liang Dong. "Scene Parsing with Deep Features and Per-Exemplar Detectors." In Lecture Notes in Electrical Engineering, 367–76. Singapore: Springer Singapore, 2017. http://dx.doi.org/10.1007/978-981-10-6445-6_40.
Повний текст джерелаТези доповідей конференцій з теми "Scene parsing"
Wang, Yu-Siang, Chenxi Liu, Xiaohui Zeng, and Alan Yuille. "Scene Graph Parsing as Dependency Parsing." In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). Stroudsburg, PA, USA: Association for Computational Linguistics, 2018. http://dx.doi.org/10.18653/v1/n18-1037.
Повний текст джерелаZhao, Hengshuang, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. "Pyramid Scene Parsing Network." In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2017. http://dx.doi.org/10.1109/cvpr.2017.660.
Повний текст джерелаZhao, Hang, Xavier Puig, Bolei Zhou, Sanja Fidler, and Antonio Torralba. "Open Vocabulary Scene Parsing." In 2017 IEEE International Conference on Computer Vision (ICCV). IEEE, 2017. http://dx.doi.org/10.1109/iccv.2017.221.
Повний текст джерелаYu, Chengcheng, Xiaobai Liu, and Song-Chun Zhu. "Single-Image 3D Scene Parsing Using Geometric Commonsense." In Twenty-Sixth International Joint Conference on Artificial Intelligence. California: International Joint Conferences on Artificial Intelligence Organization, 2017. http://dx.doi.org/10.24963/ijcai.2017/649.
Повний текст джерелаZhou, Bolei, Hang Zhao, Xavier Puig, Sanja Fidler, Adela Barriuso, and Antonio Torralba. "Scene Parsing through ADE20K Dataset." In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2017. http://dx.doi.org/10.1109/cvpr.2017.544.
Повний текст джерелаZhang, Rui, Sheng Tang, Luoqi Liu, Yongdong Zhang, Jintao Li, and Shuicheng Yan. "High Resolution Feature Recovering for Accelerating Urban Scene Parsing." In Twenty-Seventh International Joint Conference on Artificial Intelligence {IJCAI-18}. California: International Joint Conferences on Artificial Intelligence Organization, 2018. http://dx.doi.org/10.24963/ijcai.2018/161.
Повний текст джерелаZhang, Rui, Sheng Tang, Min Lin, Jintao Li, and Shuicheng Yan. "Global-residual and Local-boundary Refinement Networks for Rectifying Scene Parsing Predictions." In Twenty-Sixth International Joint Conference on Artificial Intelligence. California: International Joint Conferences on Artificial Intelligence Organization, 2017. http://dx.doi.org/10.24963/ijcai.2017/479.
Повний текст джерелаCe Liu, J. Yuen, and A. Torralba. "Nonparametric scene parsing: Label transfer via dense scene alignment." In 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2009. http://dx.doi.org/10.1109/cvprw.2009.5206536.
Повний текст джерелаLiu, Ce, Jenny Yuen, and Antonio Torralba. "Nonparametric scene parsing: Label transfer via dense scene alignment." In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPR Workshops). IEEE, 2009. http://dx.doi.org/10.1109/cvpr.2009.5206536.
Повний текст джерелаZhang, Rui, Sheng Tang, Yongdong Zhang, Jintao Li, and Shuicheng Yan. "Scale-Adaptive Convolutions for Scene Parsing." In 2017 IEEE International Conference on Computer Vision (ICCV). IEEE, 2017. http://dx.doi.org/10.1109/iccv.2017.224.
Повний текст джерела