Image-Based and Partially Categorical Annotating Approach for Pedestrian Attribute Recognition
Textual searches can be added to a person re-Identification (re-ID) surveillance system using Pedestrian Attribute Recognition (PAR). Adding a CNN-based PAR module to a re-ID model is efficient for both tasks, albeit few approaches have concentrated on modifying the data rather than the model to improve the outcomes. Without considering the cost of computing, the difficulty of deployment, or the generalizability, most contemporary multi-task methods attempt to outperform the prior methods using new models and architectures on fixed datasets. In order to examine the impact of data on the PAR result, this research suggests an image-based partially categorical attribute dataset (CA-Duke) including 36,411 images of the DukeMTMC-reID dataset for 74 pedestrian attributes. However, a systematic approach in order to determine the best location for new branches remains unclear since today's methods choose the location for additional modules on the baselines to build a multi-task network through an experimental process. In order to identify the best place to add a PAR module to a re-ID pre-trained network, this study also proposes a two-step learning method for evaluating the separability of data in the latent space via a new metric called the Separation Index (SI). Finally, extensive experiments on the attribute recognition and retrieval results indicate that comprehensive and image-based annotation can increase network proficiency by 3.31% with respect to the F1 metric. Furthermore, SI and pre-trained networks can achieve state-of-the-art performance on PAR.