Multi-object tracking (MOT) system usually consists of two tasks, object detection and re-identification (ReID). Current MOT methods tend to join detection and ReID in a single network to enhance inference speed. Such one-shot models allow joint optimization of detection and Re-ID via a shared backbone, reducing computation cost. However, the different demands of features between the two tasks in one-shot systems lead to competition in the optimization procedure. The detection task needs the features of the instances with the same class to be similar, while the ReID task needs the features of different instances to be distinguishable. Existing methods address the contradiction by disentangling the features into detection-specific and ReID-specific features. But these methods neglect the discussion of semantic interpretation of disentangling modules. In this paper, we propose a feature decoupling module, Global and Local Context-based Decoupling Module (GLCD), to disentangle features extracted by the backbone into two task-specific features. By extracting global and local contexts, the two tasks can choose different contexts by learnable parameters to enforce each self. We conduct our decoupling module into SOTA one-shot MOT method and experiments show performance improvement.
Pose estimation is a fundamental task in the field of computer vision. It contains a huge variety of sub-tasks, including 3D pose estimation, pose tracking, etc. In this paper, we propose a novel algorithm of pose estimation to determine whether it is correct of doing certain physical exercises by single-person pose-tracking and key frame extraction method. First, we use the proposed tracking compensation method to refine the output of pose estimation network. Second, we define different angles composed of human key joints as action angles to determine the standards of physical exercises. Therefore, we can have a personal physical exercise assistant and do physical exercises even without professional personal trainer around. Experiments show that our real-time method can achieve 92.0% accuracy on two kinds of physical exercise actions. It can be adapted to different applications with important significance in theory and practice.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.