Garment sewing pattern represents the intrinsic rest shape of a garment,
              and is the core for many applications like fashion design, virtual try-on,
              and digital avatars. In this work, we explore the challenging problem of
              recovering garment sewing patterns from daily photos for augmenting these
              applications. To solve the problem, we first synthesize a versatile dataset,
              named SewFactory, which consists of around 1M images and ground-truth
              sewing patterns for model training and quantitative evaluation. SewFactory
              covers a wide range of human poses, body shapes, and sewing patterns,
              and possesses realistic appearances thanks to the proposed human texture
              synthesis network. Then, we propose a two-level Transformer network
              called SewFormer, which significantly improves the sewing pattern prediction
              performance. Extensive experiments demonstrate that the proposed
              framework is effective in recovering sewing patterns and well generalizes
              to casually-taken human photos.