TIPS: Text-Induced Pose Synthesis
Document Type
Conference Article
Publication Title
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Abstract
In computer vision, human pose synthesis and transfer deal with probabilistic image generation of a person in a previously unseen pose from an already available observation of that person. Though researchers have recently proposed several methods to achieve this task, most of these techniques derive the target pose directly from the desired target image on a specific dataset, making the underlying process challenging to apply in real-world scenarios as the generation of the target image is the actual aim. In this paper, we first present the shortcomings of current pose transfer algorithms and then propose a novel text-based pose transfer technique to address those issues. We divide the problem into three independent stages: (a) text to pose representation, (b) pose refinement, and (c) pose rendering. To the best of our knowledge, this is one of the first attempts to develop a text-based pose transfer framework where we also introduce a new dataset DF-PASS, by adding descriptive pose annotations for the images of the DeepFashion dataset. The proposed method generates promising results with significant qualitative and quantitative scores in our experiments.
First Page
161
Last Page
178
DOI
10.1007/978-3-031-19839-7_10
Publication Date
1-1-2022
Recommended Citation
Roy, Prasun; Ghosh, Subhankar; Bhattacharya, Saumik; Pal, Umapada; and Blumenstein, Michael, "TIPS: Text-Induced Pose Synthesis" (2022). Conference Articles. 441.
https://digitalcommons.isical.ac.in/conf-articles/441
Comments
Open Access, Green