20 results on '"Timofte, R."'
Search Results
2. Jointly Optimized Regressors for Image Super-resolution.
- Author
-
Dai, D., Timofte, R., and Van Gool, L.
- Subjects
- *
HIGH resolution imaging , *NONLINEAR imaging sensors , *REGRESSION analysis , *MATHEMATICAL optimization , *A posteriori error analysis - Abstract
Learning regressors from low-resolution patches to high-resolution patches has shown promising results for image super-resolution. We observe that some regressors are better at dealing with certain cases, and others with different cases. In this paper, we jointly learn a collection of regressors, which collectively yield the smallest super-resolving error for all training data. After training, each training sample is associated with a label to indicate its 'best' regressor, the one yielding the smallest error. During testing, our method bases on the concept of 'adaptive selection' to select the most appropriate regressor for each input patch. We assume that similar patches can be super-resolved by the same regressor and use a fast, approximate kNN approach to transfer the labels of training patches to test patches. The method is conceptually simple and computationally efficient, yet very effective. Experiments on four datasets show that our method outperforms competing methods. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
3. Integrating Object Detection with 3D Tracking Towards a Better Driver Assistance System.
- Author
-
Prisacariu, V.A., Timofte, R., Zimmermann, K., Reid, I., and Van Gool, L.
- Published
- 2010
- Full Text
- View/download PDF
4. Multi-view traffic sign detection, recognition, and 3D localisation.
- Author
-
Timofte, R., Zimmermann, K., and Luc Van Gool
- Published
- 2009
- Full Text
- View/download PDF
5. NTIRE 2022 Spectral Recovery Challenge and Data Set
- Author
-
Boaz Arad, Radu Timofte, Rony Yahel, Nimrod Morag, Amir Bernat, Yuanhao Cai, Jing Lin, Zudi Lin, Haoqian Wang, Yulun Zhang, Hanspeter Pfister, Luc Van Gool, Shuai Liu, Yongqiang Li, Chaoyu Feng, Lei Lei, Jiaojiao Li, Songcheng Du, Chaoxiong Wu, Yihong Leng, Rui Song, Mingwei Zhang, Chongxing Song, Shuyi Zhao, Zhiqiang Lang, Wei Wei, Lei Zhang, Renwei Dian, Tianci Shan, Anjing Guo, Chengguo Feng, Jinyang Liu, Mirko Agarla, Simone Bianco, Marco Buzzelli, Luigi Celona, Raimondo Schettini, Jiang He, Yi Xiao, Jiajun Xiao, Qiangqiang Yuan, Jie Li, Liangpei Zhang, Taesung Kwon, Dohoon Ryu, Hyokyoung Bae, Hao-Hsiang Yang, Hua-En Chang, Zhi-Kai Huang, Wei-Ting Chen, Sy-Yen Kuo, Junyu Chen, Haiwei Li, Song Liu, Sabarinathan Sabarinathan, K Uma, B Sathya Bama, S. Mohamed Mansoor Roomi, Arad, B, Timofte, R, Yahel, R, Morag, N, Bernat, A, Cai, Y, Lin, J, Lin, Z, Wang, H, Zhang, Y, Pfister, H, Van Gool, L, Liu, S, Li, Y, Feng, C, Lei, L, Li, J, Du, S, Wu, C, Leng, Y, Song, R, Zhang, M, Song, C, Zhao, S, Lang, Z, Wei, W, Zhang, L, Dian, R, Shan, T, Guo, A, Liu, J, Agarla, M, Bianco, S, Buzzelli, M, Celona, L, Schettini, R, He, J, Xiao, Y, Xiao, J, Yuan, Q, Kwon, T, Ryu, D, Bae, H, Yang, H, Chang, H, Huang, Z, Chen, W, Kuo, S, Chen, J, Li, H, Sabarinathan, S, Uma, K, Bama, B, and Roomi, S
- Subjects
spectral reconstruction, spectral recovery - Abstract
This paper reviews the third biennial challenge on spectral reconstruction from RGB images, i.e., the recovery of whole-scene hyperspectral (HS) information from a 3-channel RGB image. This challenge presents the "ARAD_1K"data set: a new, larger-than-ever natural hyperspectral image data set containing 1,000 images. Challenge participants were required to recover hyper-spectral information from synthetically generated JPEG-compressed RGB images simulating capture by a known calibrated camera, operating under partially known parameters, in a setting which includes acquisition noise. The challenge was attended by 241 teams, with 60 teams com-peting in the final testing phase, 12 of which provided de-tailed descriptions of their methodology which are included in this report. The performance of these submissions is re-viewed and provided here as a gauge for the current state-of-the-art in spectral reconstruction from natural RGB images.
- Published
- 2022
6. NTIRE 2019 challenge on image enhancement: Methods and results
- Author
-
Ling Shao, Liang Lin, Flavio Piccoli, Xingguang Zhou, Dongwon Park, Syed Waqas Zamir, Lai-Kuan Wong, Greg Shakhnarovich, Cheolkon Jung, Hongzhi Zhang, Andrey Ignatov, Xiaochao Qu, Pengxu Wei, Zhiwei Zhong, Zheng Hui, Kazutoshi Akita, Jinghui Qin, Xinbo Gao, Pablo Navarrete Michelini, Wushao Wen, Jingdong Liu, Radu Timofte, Jie Liu, Jiye Liu, Salman Khan, Norimichi Ukita, Hanwen Liu, Wangmeng Zuo, Muhammad Haris, Yukai Shi, Debin Zhao, Fahad Shahbaz Khan, Pengfei Wan, Ganapathy Krishnamurthi, Xianming Liu, Se Young Chun, Simone Bianco, Tomoki Yoshida, Ting Liu, Xiumei Wang, Kai Zhang, Junjun Jiang, Claudio Cusano, John See, Nelson Chong Ngee Bow, Lishan Huang, Pengju Liu, Raimondo Schettini, Mahendra Khened, Kanti Kumari, Aditya Arora, Vikas Kumar Anand, Dan Zhu, Ignatov, A, Timofte, R, Qu, X, Zhou, X, Liu, T, Wan, P, Zamir, S, Arora, A, Khan, S, Khan, F, Shao, L, Park, D, Chun, S, Michelini, P, Liu, H, Zhu, D, Zhong, Z, Liu, X, Jiang, J, Zhao, D, Haris, M, Akita, K, Yoshida, T, Shakhnarovich, G, Ukita, N, Liu, J, Jung, C, Schettini, R, Bianco, S, Cusano, C, Piccoli, F, Liu, P, Zhang, K, Zhang, H, Zuo, W, Bow, N, Wong, L, See, J, Qin, J, Huang, L, Shi, Y, Wei, P, Wen, W, Lin, L, Hui, Z, Wang, X, Gao, X, Kumari, K, Anand, V, Khened, M, and Krishnamurthi, G
- Subjects
0209 industrial biotechnology ,Computer science ,Image quality ,Structural similarity ,business.industry ,media_common.quotation_subject ,image enhancement, image quality, tone adjustment, cameras, smartphones, task analysis, visualization, computer vision ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Visualization ,Image (mathematics) ,020901 industrial engineering & automation ,Perception ,Metric (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,Contrast (vision) ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Focus (optics) ,Image resolution ,media_common - Abstract
This paper reviews the first NTIRE challenge on perceptual image enhancement with the focus on proposed solutions and results. The participating teams were solving a real-world photo enhancement problem, where the goal was to map low-quality photos from the iPhone 3GS device to the same photos captured with Canon 70D DSLR camera. The considered problem embraced a number of computer vision subtasks, such as image denoising, image resolution and sharpness enhancement, image color/contrast/exposure adjustment, etc. The target metric used in this challenge combined PSNR and SSIM scores with solutions' perceptual results measured in the user study. The proposed solutions significantly improved baseline results, defining the state-of-the-art for practical image enhancement.
- Published
- 2019
7. NTIRE 2019 challenge on real image denoising: Methods and results
- Author
-
Kazutoshi Akita, Thomas S. Huang, Simone Zini, Raimondo Schettini, Jae-Ryun Chung, Bumjun Park, Chuan Wang, Sang-Won Lee, Seung-Won Jung, Simone Bianco, Lei Zhang, Yiyun Zhao, Yuchen Fan, Yifan Ding, Greg Shakhnarovich, Se Young Chun, Hongwei Yong, Ling Shao, Deyu Meng, Wangmeng Zuo, Chi Li, Salman Khan, Tomoki Yoshida, Chang Chen, Ding Liu, Dongwon Park, Wenyi Tang, Zhiwei Xiong, Syed Waqas Zamir, Yuqian Zhou, Norimichi Ukita, Haoqiang Fan, Seung-Wook Kim, Jue Wang, Zhiguo Cao, Yuzhi Wang, Radu Timofte, Dong-Wook Kim, Sung-Jea Ko, Fahad Shahbaz Khan, Magauiya Zhussip, Dong-Pan Lim, Seo-Won Ji, Yang Wang, Muhammad Haris, Aditya Arora, Michael S. Brown, Shakarim Soltanayev, Jiaming Liu, Qin Xu, Abdelrahman Abdelhamed, Shaofan Cai, Kai Zhang, Jechang Jeong, Chi-Hao Wu, Songhyun Yu, Yue Lu, Pengliang Tang, Abdelhamed, A, Timofte, R, Brown, M, Yu, S, Park, B, Jeong, J, Jung, S, Kim, D, Chung, J, Liu, J, Wang, Y, Wu, C, Xu, Q, Wang, C, Cai, S, Ding, Y, Fan, H, Wang, J, Zhang, K, Zuo, W, Zhussip, M, Park, D, Soltanayev, S, Chun, S, Xiong, Z, Chen, C, Haris, M, Akita, K, Yoshida, T, Shakhnarovich, G, Ukita, N, Zamir, S, Arora, A, Khan, S, Khan, F, Shao, L, Ko, S, Lim, D, Kim, S, Ji, S, Lee, S, Tang, W, Fan, Y, Zhou, Y, Liu, D, Huang, T, Meng, D, Zhang, L, Yong, H, Zhao, Y, Tang, P, Lu, Y, Schettini, R, Bianco, S, Zini, S, Li, C, and Cao, Z
- Subjects
Noise measurement ,Computer science ,business.industry ,Noise reduction ,sRGB ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,INF/01 - INFORMATICA ,02 engineering and technology ,Color space ,Real image ,Image denoising ,0202 electrical engineering, electronic engineering, information engineering ,RGB color model ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Focus (optics) ,business ,021101 geological & geomatics engineering - Abstract
This paper reviews the NTIRE 2019 challenge on real image denoising with focus on the proposed methods and their results. The challenge has two tracks for quantitatively evaluating image denoising performance in (1) the Bayer-pattern raw-RGB and (2) the standard RGB (sRGB) color spaces. The tracks had 216 and 220 registered participants, respectively. A total of 15 teams, proposing 17 methods, competed in the final phase of the challenge. The proposed methods by the 15 teams represent the current state-of-the-art performance in image denoising targeting real noisy images.
- Published
- 2019
8. Efficient End-to-End Convolutional Architecture for Point-of-Gaze Estimation.
- Author
-
Miron C, Ciubotariu G, Păsărică A, and Timofte R
- Abstract
Point-of-gaze estimation is part of a larger set of tasks aimed at improving user experience, providing business insights, or facilitating interactions with different devices. There has been a growing interest in this task, particularly due to the need for upgrades in e-meeting platforms during the pandemic when on-site activities were no longer possible for educational institutions, corporations, and other organizations. Current research advancements are focusing on more complex methodologies for data collection and task implementation, creating a gap that we intend to address with our contributions. Thus, we introduce a methodology for data acquisition that shows promise due to its nonrestrictive and straightforward nature, notably increasing the yield of collected data without compromising diversity or quality. Additionally, we present a novel and efficient convolutional neural network specifically tailored for calibration-free point-of-gaze estimation that outperforms current state-of-the-art methods on the MPIIFaceGaze dataset by a substantial margin, and sets a strong baseline on our own data.
- Published
- 2024
- Full Text
- View/download PDF
9. High-Precision Dichotomous Image Segmentation With Frequency and Scale Awareness.
- Author
-
Jiang Q, Cheng J, Wu Z, Cong R, and Timofte R
- Abstract
Dichotomous image segmentation (DIS) with rich fine-grained details within a single image is a challenging task. Despite the plausible results achieved by deep learning-based methods, most of them fail to segment generic objects when the boundary is cluttered with the background. In fact, the gradual decrease in feature map resolution during the encoding stage and the misleading texture clue may be the main issues. To handle these issues, we devise a novel frequency-and scale-aware deep neural network (FSANet) for high-precision DIS. The core of our proposed FSANet is twofold. First, a multimodality fusion (MF) module that integrates the information in spatial and frequency domains is adopted to enhance the representation capability of image features. Second, a collaborative scale fusion module (CSFM) which deviates from the traditional serial structures is introduced to maintain high resolution during the entire feature encoding stage. In the decoder side, we introduce hierarchical context fusion (HCF) and selective feature fusion (SFF) modules to infer the segmentation results from the output features of the CSFM module. We conduct extensive experiments on several benchmark datasets and compare our proposed method with existing state-of-the-art (SOTA) methods. The experimental results demonstrate that our FSANet achieves superior performance both qualitatively and quantitatively. The code will be made available at https://github.com/chasecjg/FSANet.
- Published
- 2024
- Full Text
- View/download PDF
10. Scoring facial attractiveness with deep convolutional neural networks: How training on standardized images reduces the bias of facial expressions.
- Author
-
Obwegeser D, Timofte R, Mayer C, Bornstein MM, Schätzle MA, and Patcas R
- Abstract
Objective: In many medical disciplines, facial attractiveness is part of the diagnosis, yet its scoring might be confounded by facial expressions. The intent was to apply deep convolutional neural networks (CNN) to identify how facial expressions affect facial attractiveness and to explore whether a dedicated training of the CNN is able to reduce the bias of facial expressions., Materials and Methods: Frontal facial images (n = 840) of 40 female participants (mean age 24.5 years) were taken adapting a neutral facial expression and the six universal facial expressions. Facial attractiveness was computed by means of a face detector, deep convolutional neural networks, standard support vector regression for facial beauty, visual regularized collaborative filtering and a regression technique for handling visual queries without rating history. CNN was first trained on random facial photographs from a dating website and then further trained on the Chicago Face Database (CFD) to increase its suitability to medical conditions. Both algorithms scored every image for attractiveness., Results: Facial expressions affect facial attractiveness scores significantly. Scores from CNN additionally trained on CFD had less variability between the expressions (range 54.3-60.9 compared to range: 32.6-49.5) and less variance within the scores (P ≤ .05), but also caused a shift in the ranking of the expressions' facial attractiveness., Conclusion: Facial expressions confound attractiveness scores. Training on norming images generated scores less susceptible to distortion, but more difficult to interpret. Scoring facial attractiveness based on CNN seems promising, but AI solutions must be developed on CNN trained to recognize facial expressions as distractors., (© 2024 The Author(s). Orthodontics & Craniofacial Research published by John Wiley & Sons Ltd.)
- Published
- 2024
- Full Text
- View/download PDF
11. VRT: A Video Restoration Transformer.
- Author
-
Liang J, Cao J, Fan Y, Zhang K, Ranjan R, Li Y, Timofte R, and Van Gool L
- Abstract
Video restoration aims to restore high-quality frames from low-quality frames. Different from single image restoration, video restoration generally requires to utilize temporal information from multiple adjacent but usually misaligned video frames. Existing deep methods generally tackle with this by exploiting a sliding window strategy or a recurrent architecture, which are restricted by frame-by-frame restoration. In this paper, we propose a Video Restoration Transformer (VRT) with parallel frame prediction ability. More specifically, VRT is composed of multiple scales, each of which consists of two kinds of modules: temporal reciprocal self attention (TRSA) and parallel warping. TRSA divides the video into small clips, on which reciprocal attention is applied for joint motion estimation, feature alignment and feature fusion, while self attention is used for feature extraction. To enable cross-clip interactions, the video sequence is shifted for every other layer. Besides, parallel warping is used to further fuse information from neighboring frames by parallel feature warping. Experimental results on five tasks, including video super-resolution, video deblurring, video denoising, video frame interpolation and space-time video super-resolution, demonstrate that VRT outperforms the state-of-the-art methods by large margins (up to 2.16dB) on fourteen benchmark datasets. The codes are available at https://github.com/JingyunLiang/VRT.
- Published
- 2024
- Full Text
- View/download PDF
12. PDC-Net+: Enhanced Probabilistic Dense Correspondence Network.
- Author
-
Truong P, Danelljan M, Timofte R, and Van Gool L
- Subjects
- Algorithms, Pattern Recognition, Automated methods
- Abstract
Establishing robust and accurate correspondences between a pair of images is a long-standing computer vision problem with numerous applications. While classically dominated by sparse methods, emerging dense approaches offer a compelling alternative paradigm that avoids the keypoint detection step. However, dense flow estimation is often inaccurate in the case of large displacements, occlusions, or homogeneous regions. In order to apply dense methods to real-world applications, such as pose estimation, image manipulation, or 3D reconstruction, it is therefore crucial to estimate the confidence of the predicted matches. We propose the Enhanced Probabilistic Dense Correspondence Network, PDC-Net+, capable of estimating accurate dense correspondences along with a reliable confidence map. We develop a flexible probabilistic approach that jointly learns the flow prediction and its uncertainty. In particular, we parametrize the predictive distribution as a constrained mixture model, ensuring better modelling of both accurate flow predictions and outliers. Moreover, we develop an architecture and an enhanced training strategy tailored for robust and generalizable uncertainty prediction in the context of self-supervised training. Our approach obtains state-of-the-art results on multiple challenging geometric matching and optical flow datasets. We further validate the usefulness of our probabilistic confidence estimation for the tasks of pose estimation, 3D reconstruction, image-based localization, and image retrieval.
- Published
- 2023
- Full Text
- View/download PDF
13. Learning Context-Based Nonlocal Entropy Modeling for Image Compression.
- Author
-
Li M, Zhang K, Li J, Zuo W, Timofte R, and Zhang D
- Abstract
The entropy of the codes usually serves as the rate loss in the recent learned lossy image compression methods. Precise estimation of the probabilistic distribution of the codes plays a vital role in reducing the entropy and boosting the joint rate-distortion performance. However, existing deep learning based entropy models generally assume the latent codes are statistically independent or depend on some side information or local context, which fails to take the global similarity within the context into account and thus hinders the accurate entropy estimation. To address this issue, we propose a special nonlocal operation for context modeling by employing the global similarity within the context. Specifically, due to the constraint of context, nonlocal operation is incalculable in context modeling. We exploit the relationship between the code maps produced by deep neural networks and introduce the proxy similarity functions as a workaround. Then, we combine the local and the global context via a nonlocal attention block and employ it in masked convolutional networks for entropy modeling. Taking the consideration that the width of the transforms is essential in training low distortion models, we finally produce a U-net block in the transforms to increase the width with manageable memory consumption and time complexity. Experiments on Kodak and Tecnick datasets demonstrate the priority of the proposed context-based nonlocal attention block in entropy modeling and the U-net block in low distortion situations. On the whole, our model performs favorably against the existing image compression standards and recent deep image compression models.
- Published
- 2023
- Full Text
- View/download PDF
14. Artificial intelligence in medico-dental diagnostics of the face: a narrative review of opportunities and challenges.
- Author
-
Patcas R, Bornstein MM, Schätzle MA, and Timofte R
- Subjects
- Humans, Artificial Intelligence
- Abstract
Objectives: This review aims to share the current developments of artificial intelligence (AI) solutions in the field of medico-dental diagnostics of the face. The primary focus of this review is to present the applicability of artificial neural networks (ANN) to interpret medical images, together with the associated opportunities, obstacles, and ethico-legal concerns., Material and Methods: Narrative literature review., Results: Narrative literature review., Conclusion: Curated facial images are widely available and easily accessible and are as such particularly suitable big data for ANN training. New AI solutions have the potential to change contemporary dentistry by optimizing existing processes and enriching dental care with the introduction of new tools for assessment or treatment planning. The analyses of health-related big data may also contribute to revolutionize personalized medicine through the detection of previously unknown associations. In regard to facial images, advances in medico-dental AI-based diagnostics include software solutions for the detection and classification of pathologies, for rating attractiveness and for the prediction of age or gender. In order for an ANN to be suitable for medical diagnostics of the face, the arising challenges regarding computation and management of the software are discussed, with special emphasis on the use of non-medical big data for ANN training. The legal and ethical ramifications of feeding patients' facial images to a neural network for diagnostic purposes are related to patient consent, data privacy, data security, liability, and intellectual property. Current ethico-legal regulation practices seem incapable of addressing all concerns and ensuring accountability., Clinical Significance: While this review confirms the many benefits derived from AI solutions used for the diagnosis of medical images, it highlights the evident lack of regulatory oversight, the urgent need to establish licensing protocols, and the imperative to investigate the moral quality of new norms set with the implementation of AI applications in medico-dental diagnostics., (© 2022. The Author(s).)
- Published
- 2022
- Full Text
- View/download PDF
15. Plug-and-Play Image Restoration With Deep Denoiser Prior.
- Author
-
Zhang K, Li Y, Zuo W, Zhang L, Van Gool L, and Timofte R
- Abstract
Recent works on plug-and-play image restoration have shown that a denoiser can implicitly serve as the image prior for model-based methods to solve many inverse problems. Such a property induces considerable advantages for plug-and-play image restoration (e.g., integrating the flexibility of model-based method and effectiveness of learning-based methods) when the denoiser is discriminatively learned via deep convolutional neural network (CNN) with large modeling capacity. However, while deeper and larger CNN models are rapidly gaining popularity, existing plug-and-play image restoration hinders its performance due to the lack of suitable denoiser prior. In order to push the limits of plug-and-play image restoration, we set up a benchmark deep denoiser prior by training a highly flexible and effective CNN denoiser. We then plug the deep denoiser prior as a modular part into a half quadratic splitting based iterative algorithm to solve various image restoration problems. We, meanwhile, provide a thorough analysis of parameter setting, intermediate results and empirical convergence to better understand the working mechanism. Experimental results on three representative image restoration tasks, including deblurring, super-resolution and demosaicing, demonstrate that the proposed plug-and-play image restoration with deep denoiser prior not only significantly outperforms other state-of-the-art model-based methods but also achieves competitive or even superior performance against state-of-the-art learning-based methods. The source code is available at https://github.com/cszn/DPIR.
- Published
- 2022
- Full Text
- View/download PDF
16. Using artificial intelligence to determine the influence of dental aesthetics on facial attractiveness in comparison to other facial modifications.
- Author
-
Obwegeser D, Timofte R, Mayer C, Eliades T, Bornstein MM, Schätzle MA, and Patcas R
- Subjects
- Adult, Esthetics, Dental, Face, Female, Humans, Index of Orthodontic Treatment Need, Infant, Smiling, Young Adult, Artificial Intelligence, Malocclusion therapy
- Abstract
Background: Facial aesthetics is a major motivating factor for undergoing orthodontic treatment., Objectives: To ascertain-by means of artificial intelligence (AI)-the influence of dental alignment on facial attractiveness and perceived age, compared to other modifications such as wearing glasses, earrings, or lipstick., Material and Methods: Forty volunteering females (mean age: 24.5) with near perfectly aligned upper front teeth [Aesthetic Component scale of the Index of Orthodontic Treatment Need (AC-IOTN) = 1 and Peer Assessment Rating Index (PAR Index) = 0 or 1] were photographed with a standardized pose while smiling, in the following settings (number of photographs = 960): without modifications, wearing eyeglasses, earrings, or lipstick. These pictures were taken with natural aligned dentition and with an individually manufactured crooked teeth mock-up (AC-IOTN = 8) to create the illusion of misaligned teeth. Images were assessed for attractiveness and perceived age, using AI, consisting of a face detector and deep convolutional neural networks trained on dedicated datasets for attractiveness and age prediction. Each image received an attractiveness score from 0 to 100 and one value for an age prediction. The scores were descriptively reviewed for each setting, and the facial modifications were tested statistically whether they affected the attractiveness score. The relationship between predicted age and attractiveness scores was examined with linear regression models., Results: All modifications showed a significant effect (for all: P < 0.001) on facial attractiveness. In faces with misaligned teeth, wearing eyeglasses (-17.8%) and earrings (-3.2%) had an adverse effect on facial aesthetics. Tooth alignment (+6.9%) and wearing lipstick (+7.9%) increased attractiveness. There was no relevant effect of any assessed modifications or tooth alignment on perceived age (all: <1.5 years). Mean attractiveness score declined with predicted age, except when wearing glasses, in which case attractiveness was rated higher with increasing predicted age., Conclusions: Alignment of teeth improves facial attractiveness to a similar extent than wearing lipstick, but has no discernable effect on perceived age. Wearing glasses reduces attractiveness considerably, but this effect vanishes with age., (© The Author(s) 2022. Published by Oxford University Press on behalf of the European Orthodontic Society. All rights reserved. For permissions, please email: journals.permissions@oup.com.)
- Published
- 2022
- Full Text
- View/download PDF
17. Outdoor Plant Segmentation With Deep Learning for High-Throughput Field Phenotyping on a Diverse Wheat Dataset.
- Author
-
Zenkl R, Timofte R, Kirchgessner N, Roth L, Hund A, Van Gool L, Walter A, and Aasen H
- Abstract
Robust and automated segmentation of leaves and other backgrounds is a core prerequisite of most approaches in high-throughput field phenotyping. So far, the possibilities of deep learning approaches for this purpose have not been explored adequately, partly due to a lack of publicly available, appropriate datasets. This study presents a workflow based on DeepLab v3+ and on a diverse annotated dataset of 190 RGB (350 x 350 pixels) images. Images of winter wheat plants of 76 different genotypes and developmental stages have been acquired throughout multiple years at high resolution in outdoor conditions using nadir view, encompassing a wide range of imaging conditions. Inconsistencies of human annotators in complex images have been quantified, and metadata information of camera settings has been included. The proposed approach achieves an intersection over union (IoU) of 0.77 and 0.90 for plants and soil, respectively. This outperforms the benchmarked machine learning methods which use Support Vector Classifier and/or Random Forrest. The results show that a small but carefully chosen and annotated set of images can provide a good basis for a powerful segmentation pipeline. Compared to earlier methods based on machine learning, the proposed method achieves better performance on the selected dataset in spite of using a deep learning approach with limited data. Increasing the amount of publicly available data with high human agreement on annotations and further development of deep neural network architectures will provide high potential for robust field-based plant segmentation in the near future. This, in turn, will be a cornerstone of data-driven improvement in crop breeding and agricultural practices of global benefit., Competing Interests: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest., (Copyright © 2022 Zenkl, Timofte, Kirchgessner, Roth, Hund, Van Gool, Walter and Aasen.)
- Published
- 2022
- Full Text
- View/download PDF
18. Learned Dynamic Guidance for Depth Image Reconstruction.
- Author
-
Gu S, Guo S, Zuo W, Chen Y, Timofte R, Van Gool L, and Zhang L
- Abstract
The depth images acquired by consumer depth sensors (e.g., Kinect and ToF) usually are of low resolution and insufficient quality. One natural solution is to incorporate a high resolution RGB camera and exploit the statistical correlation of its data and depth. In recent years, both optimization-based and learning-based approaches have been proposed to deal with the guided depth reconstruction problems. In this paper, we introduce a weighted analysis sparse representation (WASR) model for guided depth image enhancement, which can be considered a generalized formulation of a wide range of previous optimization-based models. We unfold the optimization by the WASR model and conduct guided depth reconstruction with dynamically changed stage-wise operations. Such a guidance strategy enables us to dynamically adjust the stage-wise operations that update the depth image, thus improving the reconstruction quality and speed. To learn the stage-wise operations in a task-driven manner, we propose two parameterizations and their corresponding methods: dynamic guidance with Gaussian RBF nonlinearity parameterization (DG-RBF) and dynamic guidance with CNN nonlinearity parameterization (DG-CNN). The network structures of the proposed DG-RBF and DG-CNN methods are designed with the the objective function of our WASR model in mind and the optimal network parameters are learned from paired training data. Such optimization-inspired network architectures enable our models to leverage the previous expertise as well as take benefit from training data. The effectiveness is validated for guided depth image super-resolution and for realistic depth image reconstruction tasks using standard benchmarks. Our DG-RBF and DG-CNN methods achieve the best quantitative results (RMSE) and better visual quality than the state-of-the-art approaches at the time of writing. The code is available at https://github.com/ShuhangGu/GuidedDepthSR.
- Published
- 2019
- Full Text
- View/download PDF
19. Facial attractiveness of cleft patients: a direct comparison between artificial-intelligence-based scoring and conventional rater groups.
- Author
-
Patcas R, Timofte R, Volokitin A, Agustsson E, Eliades T, Eichenberger M, and Bornstein MM
- Subjects
- Adult, Humans, Intelligence, Male, Young Adult, Artificial Intelligence, Face
- Abstract
Objectives: To evaluate facial attractiveness of treated cleft patients and controls by artificial intelligence (AI) and to compare these results with panel ratings performed by laypeople, orthodontists, and oral surgeons., Materials and Methods: Frontal and profile images of 20 treated left-sided cleft patients (10 males, mean age: 20.5 years) and 10 controls (5 males, mean age: 22.1 years) were evaluated for facial attractiveness with dedicated convolutional neural networks trained on >17 million ratings for attractiveness and compared to the assessments of 15 laypeople, 14 orthodontists, and 10 oral surgeons performed on a visual analogue scale (n = 2323 scorings)., Results: AI evaluation of cleft patients (mean score: 4.75 ± 1.27) was comparable to human ratings (laypeople: 4.24 ± 0.81, orthodontists: 4.82 ± 0.94, oral surgeons: 4.74 ± 0.83) and was not statistically different (all Ps ≥ 0.19). Facial attractiveness of controls was rated significantly higher by humans than AI (all Ps ≤ 0.02), which yielded lower scores than in cleft subjects. Variance was considerably large in all human rating groups when considering cases separately, and especially accentuated in the assessment of cleft patients (coefficient of variance-laypeople: 38.73 ± 9.64, orthodontists: 32.56 ± 8.21, oral surgeons: 42.19 ± 9.80)., Conclusions: AI-based results were comparable with the average scores of cleft patients seen in all three rating groups (with especially strong agreement to both professional panels) but overall lower for control cases. The variance observed in panel ratings revealed a large imprecision based on a problematic absence of unity., Implication: Current panel-based evaluations of facial attractiveness suffer from dispersion-related issues and remain practically unavailable for patients. AI could become a helpful tool to describe facial attractiveness, but the present results indicate that important adjustments are needed on AI models, to improve the interpretation of the impact of cleft features on facial attractiveness., (© The Author(s) 2019. Published by Oxford University Press on behalf of the European Orthodontic Society. All rights reserved. For permissions, please email: journals.permissions@oup.com.)
- Published
- 2019
- Full Text
- View/download PDF
20. Demosaicing Based on Directional Difference Regression and Efficient Regression Priors.
- Author
-
Wu J, Timofte R, and Van Gool L
- Subjects
- Artifacts, Colorimetry, Algorithms, Image Enhancement, Image Interpretation, Computer-Assisted
- Abstract
Color demosaicing is a key image processing step aiming to reconstruct the missing pixels from a recorded raw image. On the one hand, numerous interpolation methods focusing on spatial-spectral correlations have been proved very efficient, whereas they yield a poor image quality and strong visible artifacts. On the other hand, optimization strategies, such as learned simultaneous sparse coding and sparsity and adaptive principal component analysis-based algorithms, were shown to greatly improve image quality compared with that delivered by interpolation methods, but unfortunately are computationally heavy. In this paper, we propose efficient regression priors as a novel, fast post-processing algorithm that learns the regression priors offline from training data. We also propose an independent efficient demosaicing algorithm based on directional difference regression, and introduce its enhanced version based on fused regression. We achieve an image quality comparable to that of the state-of-the-art methods for three benchmarks, while being order(s) of magnitude faster.
- Published
- 2016
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.