31 results on '"single-view"'
Search Results
2. Enhancing Multi-view ASD Diagnosis Using Structural MRI and Pretrained CNN
- Author
-
Zemzemi, Nesrine, Hmida, Imen, Romdhane, Nadra Ben, Fendri, Emna, Li, Gang, Series Editor, Filipe, Joaquim, Series Editor, Xu, Zhiwei, Series Editor, Plácido da Silva, Hugo, editor, and Cipresso, Pietro, editor
- Published
- 2025
- Full Text
- View/download PDF
3. STNeRF: symmetric triplane neural radiance fields for novel view synthesis from single-view vehicle images: STNeRF: symmetric triplane neural radiance fields for novel...: Z. Liu et al.
- Author
-
Liu, Zhao, Fu, Zhongliang, Li, Gang, Hu, Jie, and Yang, Yang
- Abstract
This paper presents STNeRF, a method for synthesizing novel views of vehicles from single-view 2D images without the need for 3D ground truth data, such as point clouds, depth maps, CAD models, etc., as prior knowledge. A significant challenge in this task arises from the characteristics of CNNs and the utilization of local features can lead to a flattened representation of the synthesized image when training and validation with images from a single viewpoint. Many current methodologies tend to overlook local features and rely on global features throughout the entire reconstruction process, potentially resulting in the loss of fine-grained details in the synthesized image. To tackle this issue, we introduce Symmetric Triplane Neural Radiance Fields (STNeRF). STNeRF employs a triplane feature extractor with spatially aware convolution to extend 2D image features into 3D. This decouples the appearance component, which includes local features, and the shape component, which consists of global features, and utilizes them to construct a neural radiance field. These neural priors are then employed for rendering novel views. Furthermore, STNeRF leverages the symmetric properties of vehicles to liberate the appearance component from reliance on the original viewpoint and to align it with the symmetry of the target space, thereby enhancing the neural radiance field network’s ability to represent the invisible regions. The qualitative and quantitative evaluations demonstrate that STNeRF outperforms existing solutions in terms of both geometry and appearance reconstruction. More supplementary materials and the implementation code are available for access at the following link: . [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
4. Learning Reconstruction Models of Textured 3D Mesh Using StyleGAN2
- Author
-
Wang, Fei, Cao, Yangjie, Li, Zhenqiang, Li, Jie, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Huang, De-Shuang, editor, Zhang, Chuanlei, editor, and Pan, Yijie, editor
- Published
- 2024
- Full Text
- View/download PDF
5. Efficient 3D View Synthesis from Single-Image Utilizing Diffusion Priors
- Author
-
Wen, Yifan, Wang, Zitong, Li, Zhuoyuan, Wei, Dongxing, Sun, Yi, Hartmanis, Juris, Founding Editor, van Leeuwen, Jan, Series Editor, Hutchison, David, Editorial Board Member, Kanade, Takeo, Editorial Board Member, Kittler, Josef, Editorial Board Member, Kleinberg, Jon M., Editorial Board Member, Kobsa, Alfred, Series Editor, Mattern, Friedemann, Editorial Board Member, Mitchell, John C., Editorial Board Member, Naor, Moni, Editorial Board Member, Nierstrasz, Oscar, Series Editor, Pandu Rangan, C., Editorial Board Member, Sudan, Madhu, Series Editor, Terzopoulos, Demetri, Editorial Board Member, Tygar, Doug, Editorial Board Member, Weikum, Gerhard, Series Editor, Vardi, Moshe Y, Series Editor, Goos, Gerhard, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Le, Xinyi, editor, and Zhang, Zhijun, editor
- Published
- 2024
- Full Text
- View/download PDF
6. A Rapid Single-view Radar Imaging Method with Window Functions.
- Author
-
Wen Ming Yu, Yi Ting Yang, Xiao Fei Lu, Chao Yang, Zai Gao Chen, and Tie Jun Cui
- Subjects
- *
THREE-dimensional imaging , *MONOPULSE radar , *RADAR , *BISTATIC radar - Abstract
Monostatic rapid single-view radar imaging technology is a technique that employs single incidence angle and single frequency point information to implement rapid monostatic radar imaging within a small angular field. Owing to its analytical expression, this technique can substitute the traditional frequency-angle- scanning imaging in a small angular range, facilitating the rapid generation of highly realistic radar imaging data slices for complex targets and environments. This technology has been significantly applied in scatter hotspot diagnostics and target recognition. In order to achieve the windowing effect equivalent to that of frequency-angle-scanning imaging, and to enhance the scattering feature of monostatic imaging while controlling sidelobes, this paper derives analytic windowed imaging formulas for monostatic radar. It then obtains analytical expressions for various typical monostatic windowing rapid radar imaging scenarios. This enables the monostatic rapid imaging technology to maintain high efficiency in its analytical expressions while achieving the windowing effect equivalent to traditional imaging. The validity and correctness of the analytical formula and software implementation have been confirmed through 1D, 2D, and 3D imaging verifications. This technology can provide a vast amount of training data for modern radars. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Garment reconstruction from a single-view image based on pixel-aligned implicit function.
- Author
-
He, Wentao, Zhang, Ning, Song, Bingpeng, and Pan, Ruru
- Subjects
PIXELS ,CLOTHING & dress ,IMPLICIT functions ,HUMAN body ,QUANTITATIVE research - Abstract
3D garment reconstruction has a wide range of applications in apparel design, digital human body, and virtual try-on. Reconstructing 3D shapes from single-view images is a completely undefined and challenging problem. Recent single-view methods require only single-view images of static or dynamic objects and mine the potential multi-view information in single-view images by statistical, geometric, and physical prior knowledge, which is tedious to obtain. In this paper, we use an implicit function of pixel alignment represented by a neural network to correlate 2D image pixels with the corresponding 3D clothing information for end-to-end training without any relevant prior model. The qualitative and quantitative analysis of the experimental results showed that our results reduced the relative error by an average of 2.6% and the chamfer distance by 2.37% compared to the previous method. Experiments show that our model can reasonably reconstruct the 3D model of garments from their collections of single-view images. Our method can not only capture the overall geometry of the garment but also extract the tiny but important wrinkle details of the fabric. Even with low-resolution input images, our model still achieves available results. In addition, experiments show that the reconstructed 3D models of garments can be used for texture migration and virtual fitting. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. ED 2 IF 2 -Net: Learning Disentangled Deformed Implicit Fields and Enhanced Displacement Fields from Single Images Using Pyramid Vision Transformer.
- Author
-
Zhu, Xiaoqiang, Yao, Xinsheng, Zhang, Junjie, Zhu, Mengyao, You, Lihua, Yang, Xiaosong, Zhang, Jianjun, Zhao, He, and Zeng, Dan
- Subjects
TRANSFORMER models ,CONVOLUTIONAL neural networks ,DIGITAL image correlation ,COMPUTER vision ,DEEP learning ,TOPOLOGICAL fields - Abstract
There has emerged substantial research in addressing single-view 3D reconstruction and the majority of the state-of-the-art implicit methods employ CNNs as the backbone network. On the other hand, transformers have shown remarkable performance in many vision tasks. However, it is still unknown whether transformers are suitable for single-view implicit 3D reconstruction. In this paper, we propose the first end-to-end single-view 3D reconstruction network based on the Pyramid Vision Transformer (PVT), called ED 2 IF 2 -Net, which disentangles the reconstruction of an implicit field into the reconstruction of topological structures and the recovery of surface details to achieve high-fidelity shape reconstruction. ED 2 IF 2 -Net uses a Pyramid Vision Transformer encoder to extract multi-scale hierarchical local features and a global vector of the input single image, which are fed into three separate decoders. A coarse shape decoder reconstructs a coarse implicit field based on the global vector, a deformation decoder iteratively refines the coarse implicit field using the pixel-aligned local features to obtain a deformed implicit field through multiple implicit field deformation blocks (IFDBs), and a surface detail decoder predicts an enhanced displacement field using the local features with hybrid attention modules (HAMs). The final output is a fusion of the deformed implicit field and the enhanced displacement field, with four loss terms applied to reconstruct the coarse implicit field, structure details through a novel deformation loss, overall shape after fusion, and surface details via a Laplacian loss. The quantitative results obtained from the ShapeNet dataset validate the exceptional performance of ED 2 IF 2 -Net. Notably, ED 2 IF 2 -Net-L stands out as the top-performing variant, exhibiting the highest mean IoU, CD, EMD, ECD-3D, and ECD-2D scores, reaching impressive values of 61.1, 7.26, 2.51, 6.08, and 1.84, respectively. The extensive experimental evaluations consistently demonstrate the state-of-the-art capabilities of ED 2 IF 2 -Net in terms of reconstructing topological structures and recovering surface details, all while maintaining competitive inference time. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. 基于深度学习的自动睡眠分期研究综述.
- Author
-
刘颖, 储浩然, and 章浩伟
- Abstract
Copyright of Journal of Data Acquisition & Processing / Shu Ju Cai Ji Yu Chu Li is the property of Editorial Department of Journal of Nanjing University of Aeronautics & Astronautics and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2023
- Full Text
- View/download PDF
10. ARShape-Net: Single-View Image Oriented 3D Shape Reconstruction with an Adversarial Refiner
- Author
-
Xu, Hao, Bai, Jing, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Fang, Lu, editor, Chen, Yiran, editor, Zhai, Guangtao, editor, Wang, Jane, editor, Wang, Ruiping, editor, and Dong, Weisheng, editor
- Published
- 2021
- Full Text
- View/download PDF
11. HEI-Human: A Hybrid Explicit and Implicit Method for Single-View 3D Clothed Human Reconstruction
- Author
-
Liu, Leyuan, Sun, Jianchi, Gao, Yunqi, Chen, Jingying, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Ma, Huimin, editor, Wang, Liang, editor, Zhang, Changshui, editor, Wu, Fei, editor, Tan, Tieniu, editor, Wang, Yaonan, editor, Lai, Jianhuang, editor, and Zhao, Yao, editor
- Published
- 2021
- Full Text
- View/download PDF
12. 3D voxel reconstruction from single-view image based on cross-domain feature fusion.
- Author
-
Xiong, Wenjing, Huang, Fang, Zhang, Hao, and Jiang, Ming
- Subjects
- *
DATA distribution , *DEEP learning , *SUPERVISION - Abstract
The single-view 3D voxel reconstruction approach that relies on deep learning is inherently constrained by its single-input nature, which fails to account for the disparities in data distribution between the image and voxel domains. This leads to insufficient model details in the reconstruction. Currently, there is an absence of an end-to-end single-view 3D voxel reconstruction technique that effectively addresses the data distribution variances across domains while streamlining the training and inference processes. To mitigate the challenges arising from the disparities in data distribution across various domains, we introduce a novel single-view 3D voxel reconstruction model that leverages cross-domain feature fusion, termed SV3D-CDFF. SV3D-CDFF utilizes cross-domain feature clustering to mitigate the impact of data distribution disparities and employ a feature supervision method to learns voxel features. Additionally, it incorporates attention mechanism to fuse image and voxel features and utilizes residual network for 3D voxel reconstruction. Quantitative and qualitative experimental results demonstrate that SV3D-CDFF effectively integrates image and voxel features, eliminates data distribution disparities across various domains, and improves the quality of reconstructed voxel models in terms of overall structure and local detail. SV3D-CDFF outperforms existing state-of-the-arts methods in terms of the IoU and F-Score, with improvements of 0.005 and 0.045, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Learning Pairwise Inter-plane Relations for Piecewise Planar Reconstruction
- Author
-
Qian, Yiming, Furukawa, Yasutaka, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Vedaldi, Andrea, editor, Bischof, Horst, editor, Brox, Thomas, editor, and Frahm, Jan-Michael, editor
- Published
- 2020
- Full Text
- View/download PDF
14. Pix2Surf: Learning Parametric 3D Surface Models of Objects from Images
- Author
-
Lei, Jiahui, Sridhar, Srinath, Guerrero, Paul, Sung, Minhyuk, Mitra, Niloy, Guibas, Leonidas J., Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Vedaldi, Andrea, editor, Bischof, Horst, editor, Brox, Thomas, editor, and Frahm, Jan-Michael, editor
- Published
- 2020
- Full Text
- View/download PDF
15. 人工关节磨屑的显微单视图深度估计方法研究.
- Author
-
伍锐斌, 彭业萍「, 曹广忠「, 王松, and 曹树鹏「
- Subjects
STANDARD deviations ,ARTIFICIAL joints ,OPTICAL microscopes ,THREE-dimensional imaging ,FEATURE extraction - Abstract
Copyright of Lubrication Engineering (0254-0150) is the property of Editorial Office of LUBRICATION ENGINEERING and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2022
- Full Text
- View/download PDF
16. ED2IF2-Net: Learning Disentangled Deformed Implicit Fields and Enhanced Displacement Fields from Single Images Using Pyramid Vision Transformer
- Author
-
Xiaoqiang Zhu, Xinsheng Yao, Junjie Zhang, Mengyao Zhu, Lihua You, Xiaosong Yang, Jianjun Zhang, He Zhao, and Dan Zeng
- Subjects
3D reconstruction ,single-view ,deep learning ,computer vision ,transformer ,implicit field ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
There has emerged substantial research in addressing single-view 3D reconstruction and the majority of the state-of-the-art implicit methods employ CNNs as the backbone network. On the other hand, transformers have shown remarkable performance in many vision tasks. However, it is still unknown whether transformers are suitable for single-view implicit 3D reconstruction. In this paper, we propose the first end-to-end single-view 3D reconstruction network based on the Pyramid Vision Transformer (PVT), called ED2IF2-Net, which disentangles the reconstruction of an implicit field into the reconstruction of topological structures and the recovery of surface details to achieve high-fidelity shape reconstruction. ED2IF2-Net uses a Pyramid Vision Transformer encoder to extract multi-scale hierarchical local features and a global vector of the input single image, which are fed into three separate decoders. A coarse shape decoder reconstructs a coarse implicit field based on the global vector, a deformation decoder iteratively refines the coarse implicit field using the pixel-aligned local features to obtain a deformed implicit field through multiple implicit field deformation blocks (IFDBs), and a surface detail decoder predicts an enhanced displacement field using the local features with hybrid attention modules (HAMs). The final output is a fusion of the deformed implicit field and the enhanced displacement field, with four loss terms applied to reconstruct the coarse implicit field, structure details through a novel deformation loss, overall shape after fusion, and surface details via a Laplacian loss. The quantitative results obtained from the ShapeNet dataset validate the exceptional performance of ED2IF2-Net. Notably, ED2IF2-Net-L stands out as the top-performing variant, exhibiting the highest mean IoU, CD, EMD, ECD-3D, and ECD-2D scores, reaching impressive values of 61.1, 7.26, 2.51, 6.08, and 1.84, respectively. The extensive experimental evaluations consistently demonstrate the state-of-the-art capabilities of ED2IF2-Net in terms of reconstructing topological structures and recovering surface details, all while maintaining competitive inference time.
- Published
- 2023
- Full Text
- View/download PDF
17. Visual enhancement of single-view 3D point cloud reconstruction.
- Author
-
Ping, Guiju, Abolfazli Esfahani, Mahdi, Chen, Jiaying, and Wang, Han
- Subjects
- *
POINT cloud , *DEEP learning , *IMAGE reconstruction , *COMPUTER vision - Abstract
3D reconstruction from a single image is one of the core computer vision problems. Thanks to the development of deep learning, 3D reconstruction of a single image has demonstrated impressive progress in recent years. Existing researches use Chamfer distance as a loss function to guard the training of the neural network. However, the Chamfer loss will give equal weights to all points inside the 3D point clouds. It tends to sacrifice fine-grained and thin structures to avoid incurring a high loss, which will lead to visually unsatisfactory results. This paper proposes a framework that can recover a detailed three-dimensional point cloud from a single image by focusing more on boundaries (edge and corner points). Experimental results demonstrate that the proposed method outperforms existing techniques significantly, both qualitatively and quantitatively, and has fewer training parameters. [Display omitted] • Reconstructing the full 3D shape of an object from a single image. • Using a differentiable point cloud projection module to get the projected images. • Increasing visual reconstruction quality. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
18. HybridFusion: Real-Time Performance Capture Using a Single Depth Sensor and Sparse IMUs
- Author
-
Zheng, Zerong, Yu, Tao, Li, Hao, Guo, Kaiwen, Dai, Qionghai, Fang, Lu, Liu, Yebin, Hutchison, David, Series Editor, Kanade, Takeo, Series Editor, Kittler, Josef, Series Editor, Kleinberg, Jon M., Series Editor, Mattern, Friedemann, Series Editor, Mitchell, John C., Series Editor, Naor, Moni, Series Editor, Pandu Rangan, C., Series Editor, Steffen, Bernhard, Series Editor, Terzopoulos, Demetri, Series Editor, Tygar, Doug, Series Editor, Weikum, Gerhard, Series Editor, Ferrari, Vittorio, editor, Hebert, Martial, editor, Sminchisescu, Cristian, editor, and Weiss, Yair, editor
- Published
- 2018
- Full Text
- View/download PDF
19. Research on Design, Calibration and Real-Time Image Expansion Technology of Unmanned System Variable-Scale Panoramic Vision System
- Author
-
Xiaodong Guo, Zhoubo Wang, Wei Zhou, and Zhenhai Zhang
- Subjects
single-view ,catadioptric reflection ,unmanned systems ,panoramic calibration ,panoramic deployment ,Chemical technology ,TP1-1185 - Abstract
This paper summarized the research status, imaging model, systems calibration, distortion correction, and panoramic expansion of panoramic vision systems, pointed out the existing problems and put forward the prospect of future research. According to the research status of panoramic vision systems, a panoramic vision system with single viewpoint of refraction and reflection is designed. The systems had the characteristics of fast acquisition, low manufacturing cost, fixed single-view imaging, integrated imaging, and automatic switching depth of field. Based on these systems, an improved nonlinear optimization polynomial fitting method is proposed to calibrate the monocular HOVS, and the binocular HOVS is calibrated with the Aruco label. This method not only improves the robustness of the calibration results, but also simplifies the calibration process. Finally, a real-time method of panoramic map of multi-function vehicle based on vcam is proposed.
- Published
- 2021
- Full Text
- View/download PDF
20. DeepOrganNet: On-the-Fly Reconstruction and Visualization of 3D / 4D Lung Models from Single-View Projections by Deep Deformation Network.
- Author
-
Wang, Yifan, Zhong, Zichun, and Hua, Jing
- Subjects
IMAGE reconstruction ,GEOMETRIC shapes ,GEOMETRIC modeling ,LUNGS ,X-ray imaging - Abstract
This paper introduces a deep neural network based method, i.e., DeepOrganNet, to generate and visualize fully high-fidelity 3D / 4D organ geometric models from single-view medical images with complicated background in real time. Traditional 3D / 4D medical image reconstruction requires near hundreds of projections, which cost insufferable computational time and deliver undesirable high imaging / radiation dose to human subjects. Moreover, it always needs further notorious processes to segment or extract the accurate 3D organ models subsequently. The computational time and imaging dose can be reduced by decreasing the number of projections, but the reconstructed image quality is degraded accordingly. To our knowledge, there is no method directly and explicitly reconstructing multiple 3D organ meshes from a single 2D medical grayscale image on the fly. Given single-view 2D medical images, e.g., 3D / 4D-CT projections or X-ray images, our end-to-end DeepOrganNet framework can efficiently and effectively reconstruct 3D / 4D lung models with a variety of geometric shapes by learning the smooth deformation fields from multiple templates based on a trivariate tensor-product deformation technique, leveraging an informative latent descriptor extracted from input 2D images. The proposed method can guarantee to generate high-quality and high-fidelity manifold meshes for 3D / 4D lung models; while, all current deep learning based approaches on the shape reconstruction from a single image cannot. The major contributions of this work are to accurately reconstruct the 3D organ shapes from 2D single-view projection, significantly improve the procedure time to allow on-the-fly visualization, and dramatically reduce the imaging dose for human subjects. Experimental results are evaluated and compared with the traditional reconstruction method and the state-of-the-art in deep learning, by using extensive 3D and 4D examples, including both synthetic phantom and real patient datasets. The efficiency of the proposed method shows that it only needs several milliseconds to generate organ meshes with 10K vertices, which has great potential to be used in real-time image guided radiation therapy (IGRT). [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
21. Automatic single-view monocular camera calibration-based object manipulation using novel dexterous multi-fingered delta robot.
- Author
-
Kansal, Sachin and Mukherjee, Sudipto
- Subjects
- *
OBJECT manipulation , *SURFACE plates , *CAMERAS , *AUTONOMOUS robots , *ROBOTS - Abstract
The approximation of 3D geometry through single image is a particular instance of 3D reconstruction from several images. The advance information or user input should be provided to retrieve or conclude depth information. This research presents a novel automatic method to enumerated 3D affine measurements from a single perspective image. The least geometric information has been resolute through the image of the scene. The vanishing line and a vanishing point are two required information to reconstruct from an image of a scene. The affine scene structure can be reconstructed through the image of a scene. The proposed approach has many advantages; there is no need of the camera's intrinsic matrix and the explicit correlation among camera and scene (pose), no need for selecting Vx, Vy, Vz points, novel dexterous robot architecture for manipulation. In this paper, the following approaches have been implemented: (1) the three sets of vanishing points in X, Y, and Z axis; (2) the vanishing lines of the image; (3) distance among planes that parallel to the reference plane; (4) image wrapping; (5) corner detection (algorithm has been implemented in order to make the process automatic). The indigenous data set has been taken for the experiment. The results are compared with Zhang- and ArUco-based calibration. This novel approach has been used to perform tracking and manipulation of an object in real-time environment. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
22. The Research and Application of Generating 3D Urban Simulation Quickly in a Single-View Way
- Author
-
Mei, Long-Bao, Gao, Chuan, Lin, Song, editor, and Huang, Xiong, editor
- Published
- 2011
- Full Text
- View/download PDF
23. Robust Non-Rigid Motion Tracking and Surface Reconstruction Using $L_0$ Regularization.
- Author
-
Guo, Kaiwen, Liu, Yebin, Dai, Qionghai, Xu, Feng, and Wang, Yangang
- Subjects
ROBUST control ,CONSUMERS ,MATHEMATICAL regularization ,TRACKING & trailing ,COMPUTER graphics - Abstract
We present a new motion tracking technique to robustly reconstruct non-rigid geometries and motions from a single view depth input recorded by a consumer depth sensor. The idea is based on the observation that most non-rigid motions (especially human-related motions) are intrinsically involved in articulate motion subspace. To take this advantage, we propose a novel $L_0$
strategy is integrated into the available non-rigid motion tracking pipeline, and gradually extracts articulate joints information online with the tracking, which corrects the tracking errors in the results. The information of the articulate joints is used in the following tracking procedure to further improve the tracking accuracy and prevent tracking failures. Extensive experiments over complex human body motions with occlusions, facial and hand motions demonstrate that our approach substantially improves the robustness and accuracy in motion tracking. [ABSTRACT FROM PUBLISHER]- Published
- 2018
- Full Text
- View/download PDF
24. Real-Time Geometry, Albedo, and Motion Reconstruction Using a Single RGB-D Camera.
- Author
-
Guo, Kaiwen, Xu, Feng, Yu, Tao, Liu, Xiaoyang, Dai, Qionghai, and Liu, Yebin
- Subjects
REAL-time computing ,ALBEDO ,IMAGE reconstruction ,DEPTH maps (Digital image processing) ,COLOR photography - Abstract
This article proposes a real-time method that uses a single-view RGB-D input (a depth sensor integrated with a color camera) to simultaneously reconstruct a casual scene with a detailed geometry model, surface albedo, per-frame non-rigid motion, and per-frame low-frequency lighting, without requiring any template or motion priors. The key observation is that accurate scene motion can be used to integrate temporal information to recover the precise appearance, whereas the intrinsic appearance can help to establish true correspondence in the temporal domain to recover motion. Based on this observation, we first propose a shading-based scheme to leverage appearance information for motion estimation. Then, using the reconstructed motion, a volumetric albedo fusing scheme is proposed to complete and refine the intrinsic appearance of the scene by incorporating information from multiple frames. Since the two schemes are iteratively applied during recording, the reconstructed appearance and motion become increasingly more accurate. In addition to the reconstruction results, our experiments also show that additional applications can be achieved, such as relighting, albedo editing, and free-viewpoint rendering of a dynamic scene, since geometry, appearance, and motion are all reconstructed by our technique. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
25. Research on Design, Calibration and Real-Time Image Expansion Technology of Unmanned System Variable-Scale Panoramic Vision System
- Author
-
Zhenhai Zhang, Zhoubo Wang, Wei Zhou, and Xiaodong Guo
- Subjects
Technology ,Computer science ,Machine vision ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,TP1-1185 ,01 natural sciences ,Biochemistry ,Article ,Analytical Chemistry ,Nonlinear programming ,010309 optics ,Robustness (computer science) ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Calibration ,Computer vision ,Depth of field ,Electrical and Electronic Engineering ,unmanned systems ,Instrumentation ,single-view ,panoramic deployment ,Monocular ,business.industry ,Chemical technology ,Process (computing) ,panoramic calibration ,Atomic and Molecular Physics, and Optics ,Manufacturing cost ,020201 artificial intelligence & image processing ,Artificial intelligence ,catadioptric reflection ,business ,Algorithms - Abstract
This paper summarized the research status, imaging model, systems calibration, distortion correction, and panoramic expansion of panoramic vision systems, pointed out the existing problems and put forward the prospect of future research. According to the research status of panoramic vision systems, a panoramic vision system with single viewpoint of refraction and reflection is designed. The systems had the characteristics of fast acquisition, low manufacturing cost, fixed single-view imaging, integrated imaging, and automatic switching depth of field. Based on these systems, an improved nonlinear optimization polynomial fitting method is proposed to calibrate the monocular HOVS, and the binocular HOVS is calibrated with the Aruco label. This method not only improves the robustness of the calibration results, but also simplifies the calibration process. Finally, a real-time method of panoramic map of multi-function vehicle based on vcam is proposed.
- Published
- 2021
26. Random Exploration of the Procedural Space for Single-View 3D Modeling of Buildings.
- Author
-
Simon, Loic, Teboul, Olivier, Koutsourakis, Panagiotis, and Paragios, Nikos
- Subjects
- *
THREE-dimensional imaging , *PATTERN recognition systems , *COMPUTER simulation , *PIXELS , *IMAGE processing - Abstract
In this paper we tackle the problem of 3D modeling for urban environment using a modular, flexible and powerful approach driven from procedural generation. To this end, typologies of architectures are modeled through shape grammars that consist of a set of derivation rules and a set of shape/dictionary elements. Appearance (from statistical point of view with respect to the individual pixel's properties) of the dictionary elements is then learned using a set of training images. Image classifiers are trained towards recovering image support with respect to the semantics. Then, given a new image and the corresponding footprint, the modeling problem is formulated as a search of the space of shapes, that can be generated on-the-fly by deriving the grammar on the input axiom. Defining an image-based score function for the produced instances using the trained classifiers, the best rules are selected, making sure that we keep exploring the space by allowing some rules to be randomly selected. New rules are then generated by resampling around the selected rules. At the finest level, these rules define the 3D model of the building. Promising results on complex and varying architectural styles demonstrate the potential of the presented method. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
27. Visual Completion Of 3D Object Shapes From A Single View For Robotic Tasks
- Author
-
Youcef Mezouar, Juan-Antonio Corrales-Ramon, Carlos M. Mateo, Pablo Gil, Mohamed Tahoun, Omar Tahri, Institut National des Sciences Appliquées - Centre Val de Loire (INSA CVL), Institut National des Sciences Appliquées (INSA), Institut Pascal (IP), SIGMA Clermont (SIGMA Clermont)-Université Clermont Auvergne [2017-2020] (UCA [2017-2020])-Centre National de la Recherche Scientifique (CNRS), Universidad de Alicante, Universidad de Alicante. Departamento de Física, Ingeniería de Sistemas y Teoría de la Señal, Universidad de Alicante. Instituto Universitario de Investigación Informática, and Automática, Robótica y Visión Artificial
- Subjects
Visual perception ,Computer science ,3D Vision ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,stereo image processing ,Convolutional neural network ,object recognition ,Image (mathematics) ,[SPI.AUTO]Engineering Sciences [physics]/Automatic ,Object shape prediction ,convolutional neural nets ,0202 electrical engineering, electronic engineering, information engineering ,[INFO.INFO-RB]Computer Science [cs]/Robotics [cs.RO] ,Computer vision ,manipulators ,shape recognition ,single-view ,business.industry ,Deep learning ,Cognitive neuroscience of visual object recognition ,020207 software engineering ,Robotics ,3D object shape recognition ,Object (computer science) ,robot vision ,image-based robotic manipulation tasks ,configuration eye-in-hand ,3D Deep Convolutional Neural Network ,Robot ,020201 artificial intelligence & image processing ,learning (artificial intelligence) ,Artificial intelligence ,visual completion ,business ,manipulator robots ,CNN ,Ingeniería de Sistemas y Automática - Abstract
International audience; The goal of this paper is to predict 3D object shape to improve the visual perception of robots in grasping and manipulation tasks. The planning of image-based robotic manipulation tasks depends on the recognition of the object's shape. Mostly, the manipulator robots usually use a camera with configuration eye-in-hand. This fact limits the calculation of the grip on the visible part of the object. In this paper, we present a 3D Deep Convolutional Neural Network to predict the hidden parts of objects from a single-view and to accomplish recovering the complete shape of them. We have tested our proposal with both previously seen objects and novel objects from a well-known dataset.
- Published
- 2019
- Full Text
- View/download PDF
28. Research on Design, Calibration and Real-Time Image Expansion Technology of Unmanned System Variable-Scale Panoramic Vision System.
- Author
-
Guo, Xiaodong, Wang, Zhoubo, Zhou, Wei, and Zhang, Zhenhai
- Subjects
EXPERIMENTAL design ,VISION ,MONOCULARS ,INTELLIGENT transportation systems ,CAMERA calibration ,CALIBRATION - Abstract
This paper summarized the research status, imaging model, systems calibration, distortion correction, and panoramic expansion of panoramic vision systems, pointed out the existing problems and put forward the prospect of future research. According to the research status of panoramic vision systems, a panoramic vision system with single viewpoint of refraction and reflection is designed. The systems had the characteristics of fast acquisition, low manufacturing cost, fixed single-view imaging, integrated imaging, and automatic switching depth of field. Based on these systems, an improved nonlinear optimization polynomial fitting method is proposed to calibrate the monocular HOVS, and the binocular HOVS is calibrated with the Aruco label. This method not only improves the robustness of the calibration results, but also simplifies the calibration process. Finally, a real-time method of panoramic map of multi-function vehicle based on vcam is proposed. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
29. A Computer Vision Non-Contact 3D System to Improve Fingerprint Acquisition
- Author
-
Balogiannis, Georgios, Yova, Dido, Politopoulos, Konstantinos, Balogiannis, Georgios, Yova, Dido, and Politopoulos, Konstantinos
- Abstract
The fingerprint is one of the most important biometrics, with many acquisition methods developed over the years. Traditional 2D acquisition techniques produce nonlinear distortions due to the forced flattening of the finger onto a 2D surface. These random elastic deformations often introduce matching errors, making 2D techniques less reliable. Inevitably non-contact 3D capturing techniques were developed in an effort to deal with these problems. In this study we present a novel non-contact single camera 3D fingerprint reconstruction system based on fringe projection and a new model for approximating the epidermal ridges. The 3D shape of the fingerprint is reconstructed from a single 2D shading image in two steps. First the original image is decomposed into structure and texture components by an advanced Meyer algorithm. The structural component is reconstructed by a classical fringe projection technique. The textural component, containing the fingerprint information, is restored using a specialized photometric algorithm we call Cylindrical Ridge Model (CRM). CRM is a photometric algorithm that takes advantage of the axial symmetry of the ridges in order to integrate the illumination equation. The two results are combined together to form the 3D fingerprint, which is then digitally unfolded onto a 2D plane for compatibility with traditional 2D impressions. This paper describes the prototype 3D imaging system developed along with the calibration procedure, the reconstruction algorithm and the unwrapping process of the resulting 3D fingerprint, necessary for the performance evaluation of the method.
- Published
- 2016
30. Implementation and Evaluation of Verno-SP :Query Language for Massive Data Processing based on Prolog
- Author
-
Mikami, Keita, 上田, 和紀, and 上田研究室
- Subjects
Full-text Search Engine ,Prolog ,大規模データ処理 ,Single-View ,全文検索エンジン ,Massive Data Processing ,単一ビュー ,Computer science - Abstract
WWWには膨大な情報が溢れており、Vernoはそのリソースを有効に活用するための手段である、全文検索システムの一つである。Vernoは「プログラマブルな検索エンジン」をコンセプトとし、Schemeを元にした検索クエリ言語により高度な検索を可能としている。しかし、クエリの自由度が高い反面、クエリが複雑かつ大規模になってしまうという問題を抱えていた。そこで、本研究ではPrologの記述性の高さに着目し、処理系の性能測定と、Prologをベースとした検索言語「Verno-SP」の設計・実装を行った。Verno-SPの実装においては、Prologの内部DB以外に、外部のシステムも利用するが、本研究では、これら外部システムも、内部DB同様に扱えるという「データの単一ビュー」を実現した。この結果、検索クエリが大幅に簡易化・短縮化され、また実行速度面においても優れた性能を示した。, 卒業論文
- Published
- 2004
31. When Does Cotraining Work in Real Data?
- Author
-
Du, Jun, Ling, Charles X., and Zhou, Zhi-Hua
- Subjects
- *
DATA mining , *MACHINE learning , *SONAR , *UNDERWATER acoustics , *SUPERVISED learning , *ARTIFICIAL intelligence - Abstract
Cotraining, a paradigm of semisupervised learning, is promised to alleviate effectively the shortage of labeled examples in supervised learning. The standard two-view cotraining requires the data set to be described by two views of features, and previous studies have shown that cotraining works well if the two views satisfy the sufficiency and independence assumptions. In practice, however, these two assumptions are often not known or ensured (even when the two views are given). More commonly, most supervised data sets are described by one set of attributes (one view). Thus, they need be split into two views in order to apply the standard two-view cotraining. In this paper, we first propose a novel approach to empirically verify the two assumptions of cotraining given two views. Then, we design several methods to split single view data sets into two views, in order to make cotraining work reliably well. Our empirical results show that, given a whole or a large labeled training set, our view verification and splitting methods are quite effective. Unfortunately, cotraining is called for precisely when the labeled training set is small. However, given small labeled training sets, we show that the two cotraining assumptions are difficult to verify, and view splitting is unreliable. Our conclusions for cotraining's effectiveness are mixed. If two views are given, and known to satisfy the two assumptions, cotraining works well. Otherwise, based on small labeled training sets, verifying the assumptions or splitting single view into two views are unreliable; thus, it is uncertain whether the standard cotraining would work or not. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.