Статті в журналах з теми "Perceptual Quality Assessment"

Щоб переглянути інші типи публікацій з цієї теми, перейдіть за посиланням: Perceptual Quality Assessment.

Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями

Оберіть тип джерела:

Ознайомтеся з топ-50 статей у журналах для дослідження на тему "Perceptual Quality Assessment".

Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.

Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.

Переглядайте статті в журналах для різних дисциплін та оформлюйте правильно вашу бібліографію.

1

Fang, Yuming, Liping Huang, Jiebin Yan, Xuelin Liu, and Yang Liu. "Perceptual Quality Assessment of Omnidirectional Images." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 1 (June 28, 2022): 580–88. http://dx.doi.org/10.1609/aaai.v36i1.19937.

Повний текст джерела
Анотація:
Omnidirectional images, also called 360◦images, have attracted extensive attention in recent years, due to the rapid development of virtual reality (VR) technologies. During omnidirectional image processing including capture, transmission, consumption, and so on, measuring the perceptual quality of omnidirectional images is highly desired, since it plays a great role in guaranteeing the immersive quality of experience (IQoE). In this paper, we conduct a comprehensive study on the perceptual quality of omnidirectional images from both subjective and objective perspectives. Specifically, we construct the largest so far subjective omnidirectional image quality database, where we consider several key influential elements, i.e., realistic non-uniform distortion, viewing condition, and viewing behavior, from the user view. In addition to subjective quality scores, we also record head and eye movement data. Besides, we make the first attempt by using the proposed database to train a convolutional neural network (CNN) for blind omnidirectional image quality assessment. To be consistent with the human viewing behavior in the VR device, we extract viewports from each omnidirectional image and incorporate the user viewing conditions naturally in the proposed model. The proposed model is composed of two parts, including a multi-scale CNN-based feature extraction module and a perceptual quality prediction module. The feature extraction module is used to incorporate the multi-scale features, and the perceptual quality prediction module is designed to regress them to perceived quality scores. The experimental results on our database verify that the proposed model achieves the competing performance compared with the state-of-the-art methods.
Стилі APA, Harvard, Vancouver, ISO та ін.
2

Da, Pan, GuiYing Song, Ping Shi, and HaoCheng Zhang. "Perceptual quality assessment of nighttime video." Displays 70 (December 2021): 102092. http://dx.doi.org/10.1016/j.displa.2021.102092.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
3

Hamberg, Roelof, and Huib de Ridder. "Continuous assessment of perceptual image quality." Journal of the Optical Society of America A 12, no. 12 (December 1, 1995): 2573. http://dx.doi.org/10.1364/josaa.12.002573.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
4

Wang, Yinan, Andrei Chubarau, Hyunjin Yoo, Tara Akhavan, and James Clark. "Age-specific perceptual image quality assessment." Electronic Imaging 35, no. 8 (January 16, 2023): 302–1. http://dx.doi.org/10.2352/ei.2023.35.8.iqsp-302.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
5

Elloumi, Nessrine, Habiba Loukil Hadj Kacem, Nilanjan Dey, Amira S. Ashour, and Med Salim Bouhlel. "Perceptual Metrics Quality." International Journal of Service Science, Management, Engineering, and Technology 8, no. 1 (January 2017): 63–80. http://dx.doi.org/10.4018/ijssmet.2017010105.

Повний текст джерела
Анотація:
A 3D mesh can be subjected to different types of operations, such as compression, watermarking etc. Such processes lead to geometric distortions compared to the original version. In this context, quantifying the resultant modifications to the original mesh and evaluating the perceptual quality of degraded meshes become a critical issue. The perceptual 3D meshes quality is central in several applications to preserve the visual appearance of these treatments. The used metrics results have to be well correlated to the visual perception of humans. Although there are objective metrics, they do not allow the prediction of the perceptual quality, and do not include the human visual system properties. In the current work, a comparative study between the perceptual quality assessment metrics for 3D meshes was conducted. The experimental study on subjective database published by LIRIS / EPFL was used to test and to validate the results of six metrics. The results established that the Mesh Structural Distortion Measure metric achieved superior results compared to the other metrics.
Стилі APA, Harvard, Vancouver, ISO та ін.
6

Yang, Huan, Yuming Fang, and Weisi Lin. "Perceptual Quality Assessment of Screen Content Images." IEEE Transactions on Image Processing 24, no. 11 (November 2015): 4408–21. http://dx.doi.org/10.1109/tip.2015.2465145.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
7

Agudelo-Medina, Oscar A., Hernan Dario Benitez-Restrepo, Gemine Vivone, and Alan Bovik. "Perceptual Quality Assessment of Pan-Sharpened Images." Remote Sensing 11, no. 7 (April 11, 2019): 877. http://dx.doi.org/10.3390/rs11070877.

Повний текст джерела
Анотація:
Pan-sharpening (PS) is a method of fusing the spatial details of a high-resolution panchromatic (PAN) image with the spectral information of a low-resolution multi-spectral (MS) image. Visual inspection is a crucial step in the evaluation of fused products whose subjectivity renders the assessment of pansharpened data a challenging problem. Most previous research on the development of PS algorithms has only superficially addressed the issue of qualitative evaluation, generally by depicting visual representations of the fused images. Hence, it is highly desirable to be able to predict pan-sharpened image quality automatically and accurately, as it would be perceived and reported by human viewers. Such a method is indispensable for the correct evaluation of PS techniques that produce images for visual applications such as Google Earth and Microsoft Bing. Here, we propose a new image quality assessment (IQA) measure that supports the visual qualitative analysis of pansharpened outcomes by using the statistics of natural images, commonly referred to as natural scene statistics (NSS), to extract statistical regularities from PS images. Importantly, NSS are measurably modified by the presence of distortions. We analyze six PS methods in the presence of two common distortions, blur and white noise, on PAN images. Furthermore, we conducted a human study on the subjective quality of pristine and degraded PS images and created a completely blind (opinion-unaware) fused image quality analyzer. In addition, we propose an opinion-aware fused image quality analyzer, whose predictions with respect to human perceptual evaluations of pansharpened images are highly correlated.
Стилі APA, Harvard, Vancouver, ISO та ін.
8

Hu, Anzhou, Rong Zhang, Dong Yin, Yuan Chen, and Xin Zhan. "Perceptual quality assessment of SAR image compression." International Journal of Remote Sensing 34, no. 24 (October 24, 2013): 8764–88. http://dx.doi.org/10.1080/01431161.2013.846488.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
9

Chan, Kit Yan, and Ulrich Engelke. "Fuzzy regression for perceptual image quality assessment." Engineering Applications of Artificial Intelligence 43 (August 2015): 102–10. http://dx.doi.org/10.1016/j.engappai.2015.04.007.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
10

Shahriari, Y., Q. Ding, R. Fidler, M. Pelter, Y. Bai, A. Villaroman, and X. Hu. "Perceptual Image Processing Based Ecg Quality Assessment." Journal of Electrocardiology 49, no. 6 (November 2016): 937. http://dx.doi.org/10.1016/j.jelectrocard.2016.09.040.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
11

Wolfe, Virginia I., David P. Martin, and Chester I. Palmer. "Perception of Dysphonic Voice Quality by Naive Listeners." Journal of Speech, Language, and Hearing Research 43, no. 3 (June 2000): 697–705. http://dx.doi.org/10.1044/jslhr.4303.697.

Повний текст джерела
Анотація:
For clinical assessment as well as student training, there is a need for information pertaining to the perceptual dimensions of dysphonic voice. To this end, 24 naive listeners judged the similarity of 10 female and 10 male vowel samples, selected from within a narrow range of fundamental frequencies. Most of the perceptual variance for both sets of voices was associated with "degree of abnormality" as reflected by perceptual ratings as well as combined acoustic measures, based upon filtered and unfiltered signals. A second perceptual dimension for female voices was associated with high frequency noise as reflected by two acoustic measures: breathiness index (BRI) and a high-frequency power ratio. A second perceptual dimension for male voices was associated with a breathy-overtight continuum as reflected by period deviation (PDdev) and perceptual ratings of breathiness. Results are discussed in terms of perceptual training and the clinical assessment of pathological voices.
Стилі APA, Harvard, Vancouver, ISO та ін.
12

Zhai, Guangtao, Wei Sun, Xiongkuo Min, and Jiantao Zhou. "Perceptual Quality Assessment of Low-light Image Enhancement." ACM Transactions on Multimedia Computing, Communications, and Applications 17, no. 4 (November 30, 2021): 1–24. http://dx.doi.org/10.1145/3457905.

Повний текст джерела
Анотація:
Low-light image enhancement algorithms (LIEA) can light up images captured in dark or back-lighting conditions. However, LIEA may introduce various distortions such as structure damage, color shift, and noise into the enhanced images. Despite various LIEAs proposed in the literature, few efforts have been made to study the quality evaluation of low-light enhancement. In this article, we make one of the first attempts to investigate the quality assessment problem of low-light image enhancement. To facilitate the study of objective image quality assessment (IQA), we first build a large-scale low-light image enhancement quality (LIEQ) database. The LIEQ database includes 1,000 light-enhanced images, which are generated from 100 low-light images using 10 LIEAs. Rather than evaluating the quality of light-enhanced images directly, which is more difficult, we propose to use the multi-exposure fused (MEF) image and stack-based high dynamic range (HDR) image as a reference and evaluate the quality of low-light enhancement following a full-reference (FR) quality assessment routine. We observe that distortions introduced in low-light enhancement are significantly different from distortions considered in traditional image IQA databases that are well-studied, and the current state-of-the-art FR IQA models are also not suitable for evaluating their quality. Therefore, we propose a new FR low-light image enhancement quality assessment (LIEQA) index by evaluating the image quality from four aspects: luminance enhancement, color rendition, noise evaluation, and structure preserving, which have captured the most key aspects of low-light enhancement. Experimental results on the LIEQ database show that the proposed LIEQA index outperforms the state-of-the-art FR IQA models. LIEQA can act as an evaluator for various low-light enhancement algorithms and systems. To the best of our knowledge, this article is the first of its kind comprehensive low-light image enhancement quality assessment study.
Стилі APA, Harvard, Vancouver, ISO та ін.
13

Lim, Jin-Young, Ho-Seok Chang, Dong-Wook Kang, Ki-Doo Kim, and Kyeong-Hoon Jung. "No-reference Perceptual Quality Assessment of Digital Image." Journal of Broadcast Engineering 13, no. 6 (November 30, 2008): 849–58. http://dx.doi.org/10.5909/jbe.2008.13.6.849.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
14

Ma, Kede, Kai Zeng, and Zhou Wang. "Perceptual Quality Assessment for Multi-Exposure Image Fusion." IEEE Transactions on Image Processing 24, no. 11 (November 2015): 3345–56. http://dx.doi.org/10.1109/tip.2015.2442920.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
15

Wu, Yadong, Hongying Zhang, and Ran Duan. "Total Variation Based Perceptual Image Quality Assessment Modeling." Journal of Applied Mathematics 2014 (2014): 1–10. http://dx.doi.org/10.1155/2014/294870.

Повний текст джерела
Анотація:
Visual quality measure is one of the fundamental and important issues to numerous applications of image and video processing. In this paper, based on the assumption that human visual system is sensitive to image structures (edges) and image local luminance (light stimulation), we propose a new perceptual image quality assessment (PIQA) measure based on total variation (TV) model (TVPIQA) in spatial domain. The proposed measure compares TVs between a distorted image and its reference image to represent the loss of image structural information. Because of the good performance of TV model in describing edges, the proposed TVPIQA measure can illustrate image structure information very well. In addition, the energy of enclosed regions in a difference image between the reference image and its distorted image is used to measure the missing luminance information which is sensitive to human visual system. Finally, we validate the performance of TVPIQA measure with Cornell-A57, IVC, TID2008, and CSIQ databases and show that TVPIQA measure outperforms recent state-of-the-art image quality assessment measures.
Стилі APA, Harvard, Vancouver, ISO та ін.
16

Dong, Xinghui, and Huiyu Zhou. "Texture synthesis quality assessment using perceptual texture similarity." Knowledge-Based Systems 194 (April 2020): 105591. http://dx.doi.org/10.1016/j.knosys.2020.105591.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
17

Yan, Weiqing, Guanghui Yue, Yuming Fang, Hua Chen, Chang Tang, and Gangyi Jiang. "Perceptual objective quality assessment of stereoscopic stitched images." Signal Processing 172 (July 2020): 107541. http://dx.doi.org/10.1016/j.sigpro.2020.107541.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
18

Sloan, Colm, Naomi Harte, Damien Kelly, Anil C. Kokaram, and Andrew Hines. "Objective Assessment of Perceptual Audio Quality Using ViSQOLAudio." IEEE Transactions on Broadcasting 63, no. 4 (December 2017): 693–705. http://dx.doi.org/10.1109/tbc.2017.2704421.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
19

Chang, Hua-wen, Qiu-wen Zhang, Qing-gang Wu, and Yong Gan. "Perceptual image quality assessment by independent feature detector." Neurocomputing 151 (March 2015): 1142–52. http://dx.doi.org/10.1016/j.neucom.2014.04.081.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
20

Zhou Wang and Qiang Li. "Information Content Weighting for Perceptual Image Quality Assessment." IEEE Transactions on Image Processing 20, no. 5 (May 2011): 1185–98. http://dx.doi.org/10.1109/tip.2010.2092435.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
21

Chang, Hua-Wen, Hua Yang, Yong Gan, and Ming-Hui Wang. "Sparse Feature Fidelity for Perceptual Image Quality Assessment." IEEE Transactions on Image Processing 22, no. 10 (October 2013): 4007–18. http://dx.doi.org/10.1109/tip.2013.2266579.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
22

Tang, Lu, Chuangeng Tian, Leida Li, Bo Hu, Wei Yu, and Kai Xu. "Perceptual quality assessment for multimodal medical image fusion." Signal Processing: Image Communication 85 (July 2020): 115852. http://dx.doi.org/10.1016/j.image.2020.115852.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
23

Lowell, Soren Y. "The Acoustic Assessment of Voice in Continuous Speech." Perspectives on Voice and Voice Disorders 22, no. 2 (July 2012): 57–63. http://dx.doi.org/10.1044/vvd22.2.57.

Повний текст джерела
Анотація:
Acoustic measures are an essential component in the assessment of voice disorders, but the value of these measures is dependent on their relationship to perceptual voice quality and the degree to which these measures reflect the typical speaking patterns of the individual being assessed. Therefore, acoustic measures that can be accurately and reliably derived from continuous speech contexts, which are more representative of every day speaking patterns than sustained vowels, are fundamental to the assessment of voice. In this article, I review the current findings on acoustic measures that are applicable to continuous speech. I will identify spectral- and cepstral-based measures that show strong relationships to perceptual ratings of overall voice severity or relate to particular dimensions of voice quality. I also will discuss the prominence of the cepstral peak as a measure that consistently shows strong predictive capacity for perceptually rated voice severity and provides excellent discrimination of dysphonic and normal voices.
Стилі APA, Harvard, Vancouver, ISO та ін.
24

Moorthy, A. K., and A. C. Bovik. "Blind Image Quality Assessment: From Natural Scene Statistics to Perceptual Quality." IEEE Transactions on Image Processing 20, no. 12 (December 2011): 3350–64. http://dx.doi.org/10.1109/tip.2011.2147325.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
25

Zhu, Hancheng, Yong Zhou, Zhiwen Shao, Wen-Liang Du, Jiaqi Zhao, and Rui Yao. "ARET-IQA: An Aspect-Ratio-Embedded Transformer for Image Quality Assessment." Electronics 11, no. 14 (July 7, 2022): 2132. http://dx.doi.org/10.3390/electronics11142132.

Повний текст джерела
Анотація:
Image quality assessment (IQA) aims to automatically evaluate image perceptual quality by simulating the human visual system, which is an important research topic in the field of image processing and computer vision. Although existing deep-learning-based IQA models have achieved significant success, these IQA models usually require input images with a fixed size, which varies the perceptual quality of images. To this end, this paper proposes an aspect-ratio-embedded Transformer-based image quality assessment method, which can implant the adaptive aspect ratios of input images into the multihead self-attention module of the Swin Transformer. In this way, the proposed IQA model can not only relieve the variety of perceptual quality caused by size changes in input images but also leverage more global content correlations to infer image perceptual quality. Furthermore, to comprehensively capture the impact of low-level and high-level features on image quality, the proposed IQA model combines the output features of multistage Transformer blocks for jointly inferring image quality. Experimental results on multiple IQA databases show that the proposed IQA method is superior to state-of-the-art methods for assessing image technical and aesthetic quality.
Стилі APA, Harvard, Vancouver, ISO та ін.
26

Ahmed, Nisar, and Hafiz Muhammad Shahzad Asif. "Perceptual Quality Assessment of Digital Images Using Deep Features." Computing and Informatics 39, no. 3 (2020): 385–409. http://dx.doi.org/10.31577/cai_2020_3_385.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
27

Laparra, Valero, Johannes Ballé, Alexander Berardino, and Eero P. Simoncelli. "Perceptual image quality assessment using a normalized Laplacian pyramid." Electronic Imaging 2016, no. 16 (February 14, 2016): 1–6. http://dx.doi.org/10.2352/issn.2470-1173.2016.16.hvei-103.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
28

Mu, Hao, and Woon_Seng Gan. "Perceptual Quality Improvement and Assessment for Virtual Bass Systems." Journal of the Audio Engineering Society 63, no. 11 (December 2, 2015): 900–913. http://dx.doi.org/10.17743/jaes.2015.0079.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
29

Choi, Kang-Sun, Yeo-Min Yun, Jong-Woo Han, and Sung-Jea Ko. "8.4: Perceptual Quality Assessment for Motion Compensated Frame Interpolation." SID Symposium Digest of Technical Papers 41, no. 1 (2010): 102. http://dx.doi.org/10.1889/1.3499824.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
30

Kreiman, Jody, and Bruce R. Gerratt. "Perceptual Assessment of Voice Quality: Past, Present, and Future." Perspectives on Voice and Voice Disorders 20, no. 2 (July 2010): 62–67. http://dx.doi.org/10.1044/vvd20.2.62.

Повний текст джерела
Анотація:
Despite many years of research, we still do not know how to measure vocal quality. This paper reviews the history of quality assessment, describes some reasons why current approaches are unlikely to be fruitful, and proposes an alternative approach that addresses the primary difficulties with existing protocols.
Стилі APA, Harvard, Vancouver, ISO та ін.
31

Winkler, Stefan. "Issues in vision modeling for perceptual video quality assessment." Signal Processing 78, no. 2 (October 1999): 231–52. http://dx.doi.org/10.1016/s0165-1684(99)00062-6.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
32

Liu, Min, Ke Gu, Guangtao Zhai, Patrick Le Callet, and Wenjun Zhang. "Perceptual Reduced-Reference Visual Quality Assessment for Contrast Alteration." IEEE Transactions on Broadcasting 63, no. 1 (March 2017): 71–81. http://dx.doi.org/10.1109/tbc.2016.2597545.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
33

Silva, Alessandro R., and Mylène C. Q. Farias. "Perceptual quality assessment of 3D videos with stereoscopic degradations." Multimedia Tools and Applications 79, no. 1-2 (November 6, 2019): 1603–23. http://dx.doi.org/10.1007/s11042-019-08386-3.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
34

Yalman, Yildiray. "Histogram based perceptual quality assessment method for color images." Computer Standards & Interfaces 36, no. 6 (November 2014): 899–908. http://dx.doi.org/10.1016/j.csi.2014.04.002.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
35

Kuo, Wen-Hung, Po-Hung Lin, and Sheue-Ling Hwang. "A framework of perceptual quality assessment on LCD-TV." Displays 28, no. 1 (February 2007): 35–43. http://dx.doi.org/10.1016/j.displa.2006.11.005.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
36

Xia, Yingjie, Zhenguang Liu, Yan Yan, Yanxiang Chen, Luming Zhang, and Roger Zimmermann. "Media Quality Assessment by Perceptual Gaze-Shift Patterns Discovery." IEEE Transactions on Multimedia 19, no. 8 (August 2017): 1811–20. http://dx.doi.org/10.1109/tmm.2017.2679900.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
37

Mustafa, Safi, and Abdul Hameed. "Perceptual quality assessment of video using machine learning algorithm." Signal, Image and Video Processing 13, no. 8 (May 27, 2019): 1495–502. http://dx.doi.org/10.1007/s11760-019-01494-5.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
38

Zhou, Wujie, Gangyi Jiang, and Mei Yu. "New visual perceptual pooling strategy for image quality assessment." Journal of Electronics (China) 29, no. 3-4 (July 2012): 254–61. http://dx.doi.org/10.1007/s11767-012-0818-7.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
39

Farnand, Susan, Young Jang, Lark Kwon Choi, and Chuck Han. "A methodology for perceptual image quality assessment of smartphone cameras – color quality." Electronic Imaging 2017, no. 12 (January 29, 2017): 95–99. http://dx.doi.org/10.2352/issn.2470-1173.2017.12.iqsp-250.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
40

Silva, Maria Fabiana Bonfim de Lima, Sandra Madureira, Luiz Carlos Rusilo, and Zuleica Camargo. "Vocal quality assessment: methodological approach for a perceptive data analysis." Revista CEFAC 19, no. 6 (December 2017): 831–41. http://dx.doi.org/10.1590/1982-021620171961417.

Повний текст джерела
Анотація:
ABSTRACT Purpose: to present a methodological approach for interpreting perceptual judgments of vocal quality by a group of evaluators using the script Vocal Profile Analysis Scheme. Methods: a cross-sectional study based on 90 speech samples from 25 female teachers with voice disorders and/or laryngeal changes. Prior to the perceptual judgment, three perceptual tasks were performed to select samples to be presented to five evaluators using the Experiment script MFC 3.2 (software PRAAT). Next, a sequence of tests was applied, based on successive approaches of inter- and intra-evaluators’ behavior. Data were treated by statistical analysis (Cochran and Selenor tests). Results: with respect to the analysis of the evaluators' performance, it was possible to define those that presented the best results, in terms of reliability and proximity of analyses, as compared to the most experienced evaluator, excluding one. The results of the cluster analysis also allowed designing a voice quality profile of the group of speakers studied. Conclusions: the proposal of a methodological approach allowed defining evaluators whose judgments were based on phonetic knowledge, and drawing a vocal quality profile of the group of samples analyzed.
Стилі APA, Harvard, Vancouver, ISO та ін.
41

Varga, Domonkos. "No-Reference Image Quality Assessment with Global Statistical Features." Journal of Imaging 7, no. 2 (February 5, 2021): 29. http://dx.doi.org/10.3390/jimaging7020029.

Повний текст джерела
Анотація:
The perceptual quality of digital images is often deteriorated during storage, compression, and transmission. The most reliable way of assessing image quality is to ask people to provide their opinions on a number of test images. However, this is an expensive and time-consuming process which cannot be applied in real-time systems. In this study, a novel no-reference image quality assessment method is proposed. The introduced method uses a set of novel quality-aware features which globally characterizes the statistics of a given test image, such as extended local fractal dimension distribution feature, extended first digit distribution features using different domains, Bilaplacian features, image moments, and a wide variety of perceptual features. Experimental results are demonstrated on five publicly available benchmark image quality assessment databases: CSIQ, MDID, KADID-10k, LIVE In the Wild, and KonIQ-10k.
Стилі APA, Harvard, Vancouver, ISO та ін.
42

Barsties v. Latoszek, Ben, Jörg Mayer, Christopher R. Watts, and Bernhard Lehnert. "Advances in Clinical Voice Quality Analysis with VOXplot." Journal of Clinical Medicine 12, no. 14 (July 12, 2023): 4644. http://dx.doi.org/10.3390/jcm12144644.

Повний текст джерела
Анотація:
Background: The assessment of voice quality can be evaluated perceptually with standard clinical practice, also including acoustic evaluation of digital voice recordings to validate and further interpret perceptual judgments. The goal of the present study was to determine the strongest acoustic voice quality parameters for perceived hoarseness and breathiness when analyzing the sustained vowel [a:] using a new clinical acoustic tool, the VOXplot software. Methods: A total of 218 voice samples of individuals with and without voice disorders were applied to perceptual and acoustic analyses. Overall, 13 single acoustic parameters were included to determine validity aspects in relation to perceptions of hoarseness and breathiness. Results: Four single acoustic measures could be clearly associated with perceptions of hoarseness or breathiness. For hoarseness, the harmonics-to-noise ratio (HNR) and pitch perturbation quotient with a smoothing factor of five periods (PPQ5), and, for breathiness, the smoothed cepstral peak prominence (CPPS) and the glottal-to-noise excitation ratio (GNE) were shown to be highly valid, with a significant difference being demonstrated for each of the other perceptual voice quality aspects. Conclusions: Two acoustic measures, the HNR and the PPQ5, were both strongly associated with perceptions of hoarseness and were able to discriminate hoarseness from breathiness with good confidence. Two other acoustic measures, the CPPS and the GNE, were both strongly associated with perceptions of breathiness and were able to discriminate breathiness from hoarseness with good confidence.
Стилі APA, Harvard, Vancouver, ISO та ін.
43

Muschter, Evelyn, Andreas Noll, Jinting Zhao, Rania Hassen, Matti Strese, Basak Gulecyuz, Shu-Chen Li, and Eckehard Steinbach. "Perceptual Quality Assessment of Compressed Vibrotactile Signals Through Comparative Judgment." IEEE Transactions on Haptics 14, no. 2 (April 1, 2021): 291–96. http://dx.doi.org/10.1109/toh.2021.3077191.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
44

Sung, Jung-Min, Bong-Seok Choi, Bong-Yeol Choi, and Yeong-Ho Ha. "Perceptual Quality Assessment on Display based on Analytic Network Process." Journal of the Institute of Electronics and Information Engineers 51, no. 7 (July 25, 2014): 180–89. http://dx.doi.org/10.5573/ieie.2014.51.7.180.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
45

Farnand, Susan, Young Jang, Chuck Han, and Hau Hwang. "A methodology for perceptual image quality assessment of smartphone cameras." Electronic Imaging 2016, no. 13 (February 14, 2016): 1–5. http://dx.doi.org/10.2352/issn.2470-1173.2016.13.iqsp-202.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
46

Takam Tchendjou, Ghislain, and Emmanuel Simeu. "Visual Perceptual Quality Assessment Based on Blind Machine Learning Techniques." Sensors 22, no. 1 (December 28, 2021): 175. http://dx.doi.org/10.3390/s22010175.

Повний текст джерела
Анотація:
This paper presents the construction of a new objective method for estimation of visual perceiving quality. The proposal provides an assessment of image quality without the need for a reference image or a specific distortion assumption. Two main processes have been used to build our models: The first one uses deep learning with a convolutional neural network process, without any preprocessing. The second objective visual quality is computed by pooling several image features extracted from different concepts: the natural scene statistic in the spatial domain, the gradient magnitude, the Laplacian of Gaussian, as well as the spectral and spatial entropies. The features extracted from the image file are used as the input of machine learning techniques to build the models that are used to estimate the visual quality level of any image. For the machine learning training phase, two main processes are proposed: The first proposed process consists of a direct learning using all the selected features in only one training phase, named direct learning blind visual quality assessment DLBQA. The second process is an indirect learning and consists of two training phases, named indirect learning blind visual quality assessment ILBQA. This second process includes an additional phase of construction of intermediary metrics used for the construction of the prediction model. The produced models are evaluated on many benchmarks image databases as TID2013, LIVE, and LIVE in the wild image quality challenge. The experimental results demonstrate that the proposed models produce the best visual perception quality prediction, compared to the state-of-the-art models. The proposed models have been implemented on an FPGA platform to demonstrate the feasibility of integrating the proposed solution on an image sensor.
Стилі APA, Harvard, Vancouver, ISO та ін.
47

Batsi, Sophia, and Lisimachos P. Kondi. "Improved Temporal Pooling for Perceptual Video Quality Assessment Using VMAF." Electronic Imaging 2020, no. 11 (January 26, 2020): 68–1. http://dx.doi.org/10.2352/issn.2470-1173.2020.11.hvei-068.

Повний текст джерела
Анотація:
The Video Multimethod Assessment Fusion (VMAF) method, proposed by Netflix, offers an automated estimation of perceptual video quality for each frame of a video sequence. Then, the arithmetic mean of the per-frame quality measurements is taken by default, in order to obtain an estimate of the overall Quality of Experience (QoE) of the video sequence. In this paper, we validate the hypothesis that the arithmetic mean conceals the bad quality frames, leading to an overestimation of the provided quality. We also show that the Minkowski mean (appropriately parametrized) approximates well the subjectively measured QoE, providing superior Spearman Rank Correlation Coefficient (SRCC), Pearson Correlation Coefficient (PCC), and Root-Mean-Square-Error (RMSE) scores.
Стилі APA, Harvard, Vancouver, ISO та ін.
48

Guangtao Zhai, Jianfei Cai, Weisi Lin, Xiaokang Yang, Wenjun Zhang, and M. Etoh. "Cross-Dimensional Perceptual Quality Assessment for Low Bit-Rate Videos." IEEE Transactions on Multimedia 10, no. 7 (November 2008): 1316–24. http://dx.doi.org/10.1109/tmm.2008.2004910.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
49

Wang, Shiqi, Ke Gu, Kai Zeng, Zhou Wang, and Weisi Lin. "Objective Quality Assessment and Perceptual Compression of Screen Content Images." IEEE Computer Graphics and Applications 38, no. 1 (January 2018): 47–58. http://dx.doi.org/10.1109/mcg.2016.46.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
50

Oh, J., S. I. Woolley, T. N. Arvanitis, and J. N. Townend. "A multistage perceptual quality assessment for compressed digital angiogram images." IEEE Transactions on Medical Imaging 20, no. 12 (2001): 1352–61. http://dx.doi.org/10.1109/42.974930.

Повний текст джерела
Стилі APA, Harvard, Vancouver, ISO та ін.
Ми пропонуємо знижки на всі преміум-плани для авторів, чиї праці увійшли до тематичних добірок літератури. Зв'яжіться з нами, щоб отримати унікальний промокод!

До бібліографії