Persistent URL of this record https://hdl.handle.net/1887/3418406
Documents
-
- Download
- 3394171.3413673
- Publisher's Version
- open access
- Full text at publishers site
In Collections
This item can be found in the following collections:
Dual Gaussian-based variational subspace disentanglement for visible-infrared person re-identification
feature subspace, following a mixture-of-Gaussians (MoG) prior and a standard Gaussian distribution prior, respectively. Disentangling cross-modality identity-discriminable features leads to more robust retrieval for VI-ReID. To achieve efficient optimization like conventional VAE, we theoretically derive two variational inference terms for the MoG prior under the supervised setting, which not only restricts the identity-discriminable subspace so that the model...Show moreVisible-infrared person re-identification (VI-ReID) is a challenging and essential task in night-time intelligent surveillance systems. Except for the intra-modality variance that RGB-RGB person reidentification mainly overcomes, VI-ReID suffers from additional inter-modality variance caused by the inherent heterogeneous gap. To solve the problem, we present a carefully designed dual Gaussian-based variational auto-encoder (DG-VAE), which disentangles an identity-discriminable and an identity-ambiguous cross-modality
feature subspace, following a mixture-of-Gaussians (MoG) prior and a standard Gaussian distribution prior, respectively. Disentangling cross-modality identity-discriminable features leads to more robust retrieval for VI-ReID. To achieve efficient optimization like conventional VAE, we theoretically derive two variational inference terms for the MoG prior under the supervised setting, which not only restricts the identity-discriminable subspace so that the model explicitly handles the cross-modality intra-identity variance, but also enables the MoG distribution to avoid posterior collapse. Furthermore, we propose a triplet swap reconstruction (TSR) strategy to promote the above disentangling process. Extensive experiments demonstrate that our method outperforms state-of-the-art methods on two VI-ReID datasets. Codes will be available at https://github.com/TPCD/DG-VAE.Show less
- All authors
- Pu, N.; Chen, W.; Liu, Y.; Bakker, E.M.; Lew, M.S.K.
- Editor(s)
- Atrey, P.K.; Li, Z.
- Date
- 2020-10-12
- Title of host publication
- MM '20: proceedings of the 28th ACM international conference on multimedia
- Pages
- 2149 - 2158
- ISBN (print)
- 9781450379885
Conference
- Conference
- ACM MULTIMEDIA CONFERENCE 2020
- Date
- 2020-10-12 - 2020-10-16
- Location
- Seattle, United States, United States of America