25 results on '"Tian, Qi"'
Search Results
2. Multi-branch Body Region Alignment Network for Person Re-identification
- Author
-
Fang, Han, Chen, Jun, Tian, Qi, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Ro, Yong Man, editor, Cheng, Wen-Huang, editor, Kim, Junmo, editor, Chu, Wei-Ta, editor, Cui, Peng, editor, Choi, Jung-Woo, editor, Hu, Min-Chun, editor, and De Neve, Wesley, editor
- Published
- 2020
- Full Text
- View/download PDF
3. A half-precision compressive sensing framework for end-to-end person re-identification
- Author
-
Liao, Longlong, Yang, Zhibang, Liao, Qing, Li, Kenli, Li, Keqin, Liu, Jie, and Tian, Qi
- Published
- 2020
- Full Text
- View/download PDF
4. A novel two-stream saliency image fusion CNN architecture for person re-identification
- Author
-
Zhu, Fuqing, Kong, Xiangwei, Fu, Haiyan, and Tian, Qi
- Published
- 2018
- Full Text
- View/download PDF
5. Pseudo-positive regularization for deep person re-identification
- Author
-
Zhu, Fuqing, Kong, Xiangwei, Fu, Haiyan, and Tian, Qi
- Published
- 2018
- Full Text
- View/download PDF
6. Large-Scale Spatio-Temporal Person Re-Identification: Algorithms and Benchmark.
- Author
-
Shu, Xiujun, Wang, Xiao, Zang, Xianghao, Zhang, Shiliang, Chen, Yuanqi, Li, Ge, and Tian, Qi
- Subjects
ALGORITHMS ,VIDEO recording ,TASK analysis ,CAMERAS - Abstract
Person re-identification (re-ID) in the scenario with large spatial and temporal spans has not been fully explored. This fact partially occurs because existing benchmark datasets were mainly collected with limited spatial and temporal ranges, e.g., using videos recorded in a few days by cameras in a specific region of the campus. Such limited spatial and temporal ranges make it hard to simulate the difficulties of person re-ID in real scenarios. In this work, we contribute a novel Large-scale Spatio-Temporal (LaST) person re-ID dataset, including 10,862 identities with more than 228k images. Compared with existing datasets, LaST presents more challenging and high-diversity re-ID settings and significantly larger spatial and temporal ranges. For instance, each person can appear in different cities or countries, and in various time slots from day to evening, and in different seasons from spring to winter. To our best knowledge, LaST is a novel person re-ID dataset with the largest spatio-temporal ranges. Based on LaST, we verified its challenge by conducting a comprehensive performance evaluation of 14 re-ID algorithms. We further propose an easy-to-implement baseline that works well in such challenging re-ID settings. We also verified that models pre-trained on LaST can generalize well on existing datasets with short-term and cloth-changing scenarios. We expect LaST to inspire future works toward more realistic and challenging re-ID tasks. More information about the dataset is available at https://github.com/shuxjweb/last.git. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
7. Deep Shape-Aware Person Re-Identification for Overcoming Moderate Clothing Changes.
- Author
-
Chen, Jiaxing, Zheng, Wei-Shi, Yang, Qize, Meng, Jingke Meng, Hong, Richang, and Tian, Qi
- Abstract
Although person re-identification (person re-id) has advanced substantially in recent years, most methods are based on the assumption that the identities would not change clothes. This assumption might not hold in practice considering criminals who intentionally change clothes. In this work, we attempt to solve person re-id under moderate clothing change. Since the human body shape is considered as relatively more invariant under moderate clothing changes, we propose to learn a reliable shape-aware feature representation by mutually learning both colorful images and contour images. Instead of directly extracting shape features from contour images, we utilize contour feature learning as regularization and excavate more effective shape-aware feature representations from colorful images. We propose a multi-scale appearance and contour deep infomax (MAC-DIM) to maximize mutual information between colorful appearance features and contour shape features, and in this way, the extracted appearance features are constrained to be shape-aware in terms of both low-level visual properties and high-level semantics. To better model the long-range human body shape and explicitly capture contour segment relations, we introduce hierarchical graph modeling as aggregation headers, propagating structural context through graph convolutional networks (GCNs). The extensive results on benchmarks under clothing changes demonstrate the effectiveness of our shape-aware feature learning scheme. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
8. Pose-Guided Representation Learning for Person Re-Identification.
- Author
-
Li, Jianing, Zhang, Shiliang, Tian, Qi, Wang, Meng, and Gao, Wen
- Subjects
SUPERVISED learning ,IMAGE segmentation ,FEATURE extraction ,PEDESTRIANS - Abstract
The large pose variations and misalignment errors exhibited by person images significantly increase the difficulty of person Re-Identification (ReID). Existing works commonly apply extra operations like pose estimation, part segmentation, etc., to alleviate those issues and improve the robustness of pedestrian representations. While boosting the ReID accuracy, those operations introduce considerable computational overheads and make the deep models complex and hard to tune. To chase a more efficient solution, we propose a Part-Guided Representation (PGR) composed of Pose Invariant Feature (PIF) and Local Descriptive Feature (LDF), respectively. We call PGR “Part-Guided” because it is trained and supervised by local part cues. Specifically, PIF approximates a pose invariant representation inferred by pose estimation and pose normalization. LDF focuses on discriminative body parts by approximating a representation learned with body region segmentation. In this way, extra pose extraction is only introduced during the training stage to supervise the learning of PGR, but is not required during the testing stage for feature extraction. Extensive comparisons with recent works on five widely used datasets demonstrate the competitive accuracy and efficiency of PGR. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
9. Camera-Based Batch Normalization: An Effective Distribution Alignment Method for Person Re-Identification.
- Author
-
Zhuang, Zijie, Wei, Longhui, Xie, Lingxi, Ai, Haizhou, and Tian, Qi
- Subjects
CAMERAS ,FEATURE extraction - Abstract
Person re-identification (ReID) aims at matching identities across disjoint cameras. Its fundamental difficulty lies in associating images across individual cameras, where a key clue, i.e., identity appearance, is prone to the environmental factors of cameras and, consequently, subject to distinct image distributions due to the environmental differences between cameras. To associate images from training cameras, ReID methods strongly demand expensive inter-camera annotations for learning the relations between the distribution of these cameras, yet trained models are still not guaranteed to transfer well to unseen cameras. This problem significantly limits the application of ReID. This paper rethinks the working mechanism of conventional ReID approaches and puts forward a new solution. With an effective operator named Camera-based Batch Normalization (CBN), we guarantee an invariant input distribution independent of all cameras. Thus, the training and testing procedures are always conducted under the same input distribution. This alignment brings three benefits. First, ReID models enjoy better abilities to generalize across testing scenarios with unseen cameras and transfer across multiple training sets. Second, it makes better use of intra-camera annotations, which have been undervalued before due to the lack of cross-camera information. Ideally, the cost of inter-camera annotations can be largely reduced. Third, cross-modality tasks can be better defined through aligning visible/infrared cameras’ distributions. Experiments on a wide range of ReID tasks demonstrate the effectiveness of our approach. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
10. An End-to-End Foreground-Aware Network for Person Re-Identification.
- Author
-
Liu, Yiheng, Zhou, Wengang, Liu, Jianzhuang, Qi, Guo-Jun, Tian, Qi, and Li, Houqiang
- Subjects
PEDESTRIANS ,IDENTIFICATION ,FEATURE extraction - Abstract
Person re-identification is a crucial task of identifying pedestrians of interest across multiple surveillance camera views. For person re-identification, a pedestrian is usually represented with features extracted from a rectangular image region that inevitably contains the scene background, which incurs ambiguity to distinguish different pedestrians and degrades the accuracy. Thus, we propose an end-to-end foreground-aware network to discriminate against the foreground from the background by learning a soft mask for person re-identification. In our method, in addition to the pedestrian ID as supervision for the foreground, we introduce the camera ID of each pedestrian image for background modeling. The foreground branch and the background branch are optimized collaboratively. By presenting a target attention loss, the pedestrian features extracted from the foreground branch become more insensitive to backgrounds, which greatly reduces the negative impact of changing backgrounds on pedestrian matching across different camera views. Notably, in contrast to existing methods, our approach does not require an additional dataset to train a human landmark detector or a segmentation model for locating the background regions. The experimental results conducted on three challenging datasets, i.e., Market-1501, DukeMTMC-reID, and MSMT17, demonstrate the effectiveness of our approach. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
11. Learning to Align via Wasserstein for Person Re-Identification.
- Author
-
Zhang, Zhizhong, Xie, Yuan, Li, Ding, Zhang, Wensheng, and Tian, Qi
- Subjects
CONVOLUTIONAL neural networks ,INSTRUCTIONAL systems - Abstract
Existing successful person re-identification (Re-ID) models often employ the part-level representation to extract the fine-grained information, but commonly use the loss that is particularly designed for global features, ignoring the relationship between semantic parts. In this paper, we present a novel triplet loss that emphasizes the salient parts and also takes the consideration of alignment. This loss is based on the crossing-bing matching metric that also known as Wasserstein Distance. It measures how much effort it would take to move the embeddings of local features to align two distributions, such that it is able to find an optimal transport matrix to re-weight the distance of different local parts. The distributions in support of local parts is produced via a new attention mechanism, which is calculated by the inner product between high-level global feature and local features, representing the importance of different semantic parts w.r.t. identification. We show that the obtained optimal transport matrix can not only distinguish the relevant and misleading parts, and hence assign different weights to them, but also rectify the original distance according to the learned distributions, resulting in an elegant solution for the mis-alignment issue. Besides, the proposed method is easily implemented in most Re-ID learning system with end-to-end training style, and can obviously improve their performance. Extensive experiments and comparisons with recent Re-ID methods manifest the competitive performance of our method. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
12. Tensor Multi-Task Learning for Person Re-Identification.
- Author
-
Zhang, Zhizhong, Xie, Yuan, Zhang, Wensheng, Tang, Yongqiang, and Tian, Qi
- Subjects
VECTOR spaces ,UNIFORM spaces ,CAMERAS ,LEARNING - Abstract
This article presents a tensor multi-task model for person re-identification (Re-ID). Due to discrepancy among cameras, our approach regards Re-ID from multiple cameras as different but related classification tasks, each task corresponding to a specific camera. In each task, we distinguish the person identity as a one-vs-all linear classification problem, where one classifier is associated with a specific person. By constructing all classifiers into a task-specific projection matrix, the proposed method could utilize all the matrices to form a tensor structure, and jointly train all the tasks in a uniform tensor space. In this space, by assuming the features of the same person under different cameras are generated from a latent subspace, and different identities under the same perspective share similar patterns, the high-order correlations, not only across different tasks but also within a certain task, can be captured by utilizing a new type of low-rank tensor constraint. Therefore, the learned classifiers transform the original feature vector into the latent space, where feature distributions across cameras can be well-aligned. Moreover, this model can be incorporated into multiple visual features to boost the performance, and easily extended to the unsupervised setting. Extensive experiments and comparisons with recent Re-ID methods manifest the competitive performance of our method. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
13. A Survey of Open-World Person Re-Identification.
- Author
-
Leng, Qingming, Ye, Mang, and Tian, Qi
- Subjects
PATTERN recognition systems ,WORK design - Abstract
Person re-identification (re-ID) has been a popular topic in computer vision and pattern recognition communities for a decade. Several important milestones such as metric-based and deeply-learned re-ID in recent years have promoted this topic. However, most existing re-ID works are designed for closed-world scenarios rather than realistic open-world settings, which limits the practical application of the re-ID technique. On one hand, the performance of the latest re-ID methods has surpassed the human-level performance on several commonly used benchmarks (e.g., Market1501 and CUHK03), which are collected from closed-world scenarios. On the other hand, open-world tasks that are less developed and more challenging have received increasing attention in the re-ID community. Therefore, this paper starts the first attempt to analyze the trends of open-world re-ID and summarizes them from both narrow and generalized perspectives. In the narrow perspective, open-world re-ID is regarded as person verification (i.e., open-set re-ID) instead of person identification, that is, the query person may not occur in the gallery set. In the generalized perspective, application-driven methods that are designed for specific applications are defined as generalized open-world re-ID. Their settings are usually close to realistic application requirements. Specifically, this survey mainly includes the following four points for open-world re-ID: 1) analyzing the discrepancies between closed- and open-world scenarios; 2) describing the developments of existing open-set re-ID works and their limitations; 3) introducing specific application-driven works from three aspects, namely, raw data, practical procedure, and efficiency; and 4) summarizing the state-of-the-art methods and future directions for open-world re-ID. This survey on open-world re-ID provides a guidance for improving the usability of re-ID technique in practical applications. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
14. Effective Image Retrieval via Multilinear Multi-Index Fusion.
- Author
-
Zhang, Zhizhong, Xie, Yuan, Zhang, Wensheng, and Tian, Qi
- Abstract
Multi-index fusion has demonstrated impressive performances in the retrieval task by integrating different visual representations in a unified framework. However, previous works mainly consider propagating similarities via a neighbor structure, ignoring the high-order information among different visual representations. In this paper, we propose a new multi-index fusion scheme for image retrieval. By formulating this procedure as a multilinear-based optimization problem, the complementary information hidden in different indexes can be explored more thoroughly. Specifically, we first build our multiple indexes from various visual representations. Then, a so-called index-specific functional matrix, which aims to propagate similarities, is introduced to update the original index. The functional matrices are then optimized in a unified tensor space to achieve a refinement, such that the relevant images can be pushed closer. The optimization problem can be efficiently solved by the augmented Lagrangian method with a theoretical convergence guarantee. Unlike the traditional multi-index fusion scheme, our approach embeds the multi-index subspace structure into the new indexes with sparse constraint and, thus, it has little additional memory consumption in the online query stage. Experimental evaluation on three benchmark datasets reveals that the proposed approach achieves state-of-the-art performance, that is, N-score 3.94 on UKBench, mAP 94.1% on Holiday, and 62.39% on Market-1501. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
15. An Adaptive Multi-Projection Metric Learning for Person Re-Identification Across Non-Overlapping Cameras.
- Author
-
Hu, Hai-Miao, Fang, Wen, Li, Bo, and Tian, Qi
- Subjects
CAMERAS ,ADAPTIVE optics ,IMAGE color analysis ,SOFTWARE measurement - Abstract
Person re-identification is one of the most important and challenging problems in video analytics systems; it aims to match people across non-overlapping camera views. For person re-identification, metric learning is introduced to improve the performance by providing a metric adapted for cross-view matching. The essence of metric learning is to search for an optimal projection matrix to project the original features into a new feature space. However, most existing metric learning methods overlook the inconsistency of feature distributions in multiple cameras. In this paper, we propose a multi-projection metric learning (MPML) method to overcome the inconsistency among multiple cameras in person re-identification. Our solution is to jointly learn multiple projection matrices using paired samples from different cameras to project features from different cameras into a common feature space. To make our method adaptive to newly added cameras without affecting the learned projection matrices, we further propose an adaptive MPML method, which can learn new camera projection matrices without having to update any of the obtained projection matrices. The proposed methods are evaluated on four major person re-identification data sets, with comprehensive experiments showing the effectiveness of the proposed methods and notable improvements over the state-of-the–art approaches. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
16. Deep Representation Learning With Part Loss for Person Re-Identification.
- Author
-
Yao, Hantao, Zhang, Shiliang, Hong, Richang, Zhang, Yongdong, Xu, Changsheng, and Tian, Qi
- Subjects
DEEP learning ,HUMAN body ,IMAGE representation ,THIRD parties (Law) ,FEATURE extraction - Abstract
Learning discriminative representations for unseen person images is critical for person re-identification (ReID). Most of the current approaches learn deep representations in classification tasks, which essentially minimize the empirical classification risk on the training set. As shown in our experiments, such representations easily get over-fitted on a discriminative human body part on the training set. To gain the discriminative power on unseen person images, we propose a deep representation learning procedure named part loss network, to minimize both the empirical classification risk on training person images and the representation learning risk on unseen person images. The representation learning risk is evaluated by the proposed part loss, which automatically detects human body parts and computes the person classification loss on each part separately. Compared with traditional global classification loss, simultaneously considering part loss enforces the deep network to learn representations for different body parts and gain the discriminative power on unseen persons. Experimental results on three person ReID datasets, i.e., Market1501, CUHK03, and VIPeR, show that our representation outperforms existing deep representations. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
17. Multi-Task Learning with Low Rank Attribute Embedding for Multi-Camera Person Re-Identification.
- Author
-
Su, Chi, Yang, Fan, Zhang, Shiliang, Tian, Qi, Davis, Larry Steven, and Gao, Wen
- Subjects
CAMERAS ,EMBEDDINGS (Mathematics) ,MATHEMATICAL optimization ,LINEAR programming ,DATA visualization ,MACHINE learning - Abstract
We propose Multi-Task Learning with Low Rank Attribute Embedding (MTL-LORAE) to address the problem of person re-identification on multi-cameras. Re-identifications on different cameras are considered as related tasks, which allows the shared information among different tasks to be explored to improve the re-identification accuracy. The MTL-LORAE framework integrates low-level features with mid-level attributes as the descriptions for persons. To improve the accuracy of such description, we introduce the low-rank attribute embedding, which maps original binary attributes into a continuous space utilizing the correlative relationship between each pair of attributes. In this way, inaccurate attributes are rectified and missing attributes are recovered. The resulting objective function is constructed with an attribute embedding error and a quadratic loss concerning class labels. It is solved by an alternating optimization strategy. The proposed MTL-LORAE is tested on four datasets and is validated to outperform the existing methods with significant margins. [ABSTRACT FROM PUBLISHER]
- Published
- 2018
- Full Text
- View/download PDF
18. Diverse part attentive network for video-based person re-identification.
- Author
-
Shu, Xiujun, Li, Ge, Wei, Longhui, Zhong, Jia-Xing, Zang, Xianghao, Zhang, Shiliang, Wang, Yaowei, Liang, Yongsheng, and Tian, Qi
- Subjects
- *
HUMAN body , *COMPUTATIONAL complexity - Abstract
• We propose a lightweight attention mechanism to exploit diverse parts of human bodies for addressing visual variations. • We propose an effective framework for video-based person re-identification. • We conduct extensive experiments on three popular benchmarks for demonstrating the effectiveness of our proposed method. Attention mechanisms have achieved success in video-based person re-identification (re-ID). However, current global attentions tend to focus on the most salient parts, e.g., clothes, and ignore other subtle but valuable cues, e.g., hair, bag, and shoes. They still do not make full use of valuable information from diverse parts of human bodies. To tackle this issue, we propose a Diverse Part Attentive Network (DPAN) to exploit discriminative and diverse body cues. The framework consists of two modules: spatial diverse part attention and temporal diverse part attention. The spatial module utilizes channel grouping to exploit diverse parts of human bodies including salient and subtle parts. The temporal module aims to learn diverse weights for fusing learned features. Besides, this framework is lightweight, which introduces marginal parameters and computational complexities. Extensive experiments were conducted on three popular benchmarks, i.e. iLIDS-VID, PRID2011 and MARS. Our method achieves competitive performance on these datasets compared with state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
19. Attribute Discovery for Person Re-Identification
- Author
-
Umeda, Takayuki, Sun, Yongqing, Irie, Go, Sudo, Kyoko, Kinebuchi, Tetsuya, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Tian, Qi, editor, Sebe, Nicu, editor, Qi, Guo-Jun, editor, Huet, Benoit, editor, Hong, Richang, editor, and Liu, Xueliang, editor
- Published
- 2016
- Full Text
- View/download PDF
20. Spatial Constrained Fine-Grained Color Name for Person Re-identification
- Author
-
Yang, Yang, Yang, Yuhong, Ye, Mang, Huang, Wenxin, Wang, Zheng, Liang, Chao, Yao, Lei, Zhang, Chunjie, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Tian, Qi, editor, Sebe, Nicu, editor, Qi, Guo-Jun, editor, Huet, Benoit, editor, Hong, Richang, editor, and Liu, Xueliang, editor
- Published
- 2016
- Full Text
- View/download PDF
21. Camera Network Based Person Re-identification by Leveraging Spatial-Temporal Constraint and Multiple Cameras Relations
- Author
-
Huang, Wenxin, Hu, Ruimin, Liang, Chao, Yu, Yi, Wang, Zheng, Zhong, Xian, Zhang, Chunjie, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Tian, Qi, editor, Sebe, Nicu, editor, Qi, Guo-Jun, editor, Huet, Benoit, editor, Hong, Richang, editor, and Liu, Xueliang, editor
- Published
- 2016
- Full Text
- View/download PDF
22. Multi-type attributes driven multi-camera person re-identification.
- Author
-
Su, Chi, Zhang, Shiliang, Xing, Junliang, Gao, Wen, and Tian, Qi
- Subjects
- *
ARTIFICIAL neural networks , *SUPERVISED learning , *ROBUST control , *CONTEXTUAL analysis , *GENERALIZABILITY theory - Abstract
One of the major challenges in person Re-Identification (ReID) is the inconsistent visual appearance of a person. Current works on visual feature and distance metric learning have achieved significant achievements, but still suffer from the limited robustness to pose variations, viewpoint changes, etc ., and the high computational complexity. This makes person ReID among multiple cameras still challenging. This work is motivated to learn mid-level human attributes which are robust to visual appearance variations and could be used as efficient features for person matching. We propose a weakly supervised multi-type attribute learning framework which considers the contextual cues among attributes and progressively boosts the accuracy of attributes only using a limited number of labeled data. Specifically, this framework involves a three-stage training. A deep Convolutional Neural Network (dCNN) is first trained on an independent dataset labeled with attributes. Then it is fine-tuned on another dataset only labeled with person IDs using our defined triplet loss. Finally, the updated dCNN predicts attribute labels for the target dataset, which is combined with the independent dataset for the final round of fine-tuning. The predicted attributes, namely deep attributes exhibit promising generalization ability across different datasets. By directly using the deep attributes with simple Cosine distance, we have obtained competitive accuracy on four person ReID datasets. Experiments also show that a simple distance metric learning modular further boosts our method, making it outperform many recent works. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
23. Attributes driven tracklet-to-tracklet person re-identification using latent prototypes space mapping.
- Author
-
Su, Chi, Zhang, Shiliang, Yang, Fan, Zhang, Guangxiao, Tian, Qi, Gao, Wen, and Davis, Larry S.
- Subjects
- *
BIOMETRIC identification , *PROTOTYPES , *CARTOGRAPHY , *IMAGE processing , *COMPUTER algorithms , *SEMANTIC networks (Information theory) - Abstract
Most of current person re-identification works identify a person by matching his/her probe image against a galley images set. One feasible way to improve the identification accuracy is the multi-shot re-identification, where the probe includes a small set of images rather than a single image. In this paper, we study the tracklet-to-tracklet identification, where both the probe and the target dataset are composed of small sets of sequential images, i.e., tracklets. To solve this problem and make our algorithm robust under multi-camera setting, we take full advantage of low-level features, attributes and inter-attribute correlations at the same time. Attributes are expected to offer semantic information complementary to low-level features. In order to discover the correlations among attributes, a novel discriminative model is proposed to exploit low-level features and map attributes to a discriminative latent prototypes space. An alternating optimization procedure is designed to perform the learning process. We also devise a number of voting schemes to total up matching scores from images to tracklets. Experiments on four public datasets show that our approach achieves a consistently better performance than existing person re-identification methods. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
24. Towards generalizable person re-identification with a bi-stream generative model.
- Author
-
Xu, Xin, Liu, Wei, Wang, Zheng, Hu, Ruimin, and Tian, Qi
- Subjects
- *
PEDESTRIANS , *HUMAN body , *LINEAR network coding , *POSE estimation (Computer vision) - Abstract
• We decouple the difficulties affecting the person re-identification task into the Camera-Camera (CC) problem and the Camera-Person (CP) problem. • We propose a bi-stream generative model for solving the CC and CP problems separately, with promising results. • We design a part-weighted loss based on the unbalanced number of human body parts in the dataset to guide the model to focus on the more important parts. Generalizable person re-identification (re-ID) has attracted growing attention due to its powerful adaptation capability in the unseen data domain. However, existing solutions often neglect either crossing cameras (e.g., illumination and resolution differences) or pedestrian misalignments (e.g., viewpoint and pose discrepancies), which easily leads to poor generalization capability when adapted to the new domain. In this paper, we formulate these difficulties as: 1) Camera-Camera (CC) problem, which denotes the various human appearance changes caused by different cameras; 2) Camera-Person (CP) problem, which indicates the pedestrian misalignments caused by the same identity person under different camera viewpoints or changing pose. To solve the above issues, we propose a Bi-stream Generative Model (BGM) to learn the fine-grained representations fused with camera-invariant global feature and pedestrian-aligned local feature, which contains an encoding network and two stream decoding sub-network. Guided by original pedestrian images, one stream is employed to learn a camera-invariant global feature for the CC problem via filtering cross-camera interference factors. For the CP problem, another stream learns a pedestrian-aligned local feature for pedestrian alignment using information-complete densely semantically aligned part maps. Moreover, a part-weighted loss function is presented to reduce the influence of missing parts on pedestrian alignment. Extensive experiments demonstrate that our method outperforms the state-of-the-art methods on the large-scale generalizable re-ID benchmarks, involving domain generalization setting and cross-domain setting. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
25. 3D-GAT: 3D-Guided adversarial transform network for person re-identification in unseen domains.
- Author
-
Zhang, Hengheng, Li, Ying, Zhuang, Zijie, Xie, Lingxi, and Tian, Qi
- Subjects
- *
CAMERAS , *ALGORITHMS , *TRANSLATIONS , *WITNESSES , *ABILITY , *STEREO vision (Computer science) - Abstract
Person Re-identification (ReID) has witnessed remarkable improvements in the past couple of years. However, its applications in real-world scenarios are limited by the disparity among different cameras and datasets. In general, it remains challenging to generalize ReID algorithms from one domain to another, especially when the target domain is unknown. To solve this issue, we develop a 3D-guided adversarial transform (3D-GAT) network which explores the transfer ability of source training data to facilitate learning domain-independent knowledge. Being aware of a 3D model and human poses, 3D-GAT makes use of image-to-image translation to synthesize person images in different conditions whilst preserving features for identification as much as possible. With these augmented training data, it is easier for ReID approaches to perceive how a person can appear differently under varying viewpoints and poses, most of which are not seen in the training data, and thus achieve higher ReID accuracy especially in an unknown domain. Extensive experiments conducted on Market-1501, DukeMTMC-reID and CUHK03 demonstrate the effectiveness of our proposed approach, which is competitive to the baseline models in the original dataset and sets the new state-of-the-art in direct transfer to other datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.