ETCH-X
Robustify Body Fitting to Any Clothed Human Scans with Composable Synthetic Data

1Zhejiang University 2Westlake University 3Shanghai Innovation Institute 4Fudan University 5Nanjing University
Project Teaser Image

Strengths of ETCH-X. While NICP, which uses implicit dense correspondence but lacks tightness-aware undressing, consistently produces overweight bodies from clothed scans (A), ETCH, with tightness-aware undressing but sparse markers, fails to capture detailed body parts such as hands and face (B), and struggles with partial inputs due to missing markers (C). In contrast, our ETCH-X combines the strengths of both approaches, achieving robust and expressive fitting across diverse clothing, poses, and levels of input completeness (D).

Abstract

Human body fitting, which aligns parametric body models such as SMPL to raw 3D point clouds of clothed humans, serves as a crucial first step for downstream tasks like animation and texturing. An effective fitting method should be both locally expressive — capturing fine details such as hands and facial features — and globally robust to handle real-world challenges, including clothing dynamics, pose variations, and noisy or partial inputs. Existing approaches typically excel in only one aspect, lacking an all-in-one solution. We upgrade ETCH to ETCH-X, which leverages a tightness-aware fitting paradigm to filter out clothing dynamics ("undress"), extends expressiveness with SMPL-X, and replaces explicit sparse markers (which are highly sensitive to partial data) with implicit dense correspondences ("dense fit") for more robust and fine-grained body fitting. Our disentangled "undress" and "dense fit" modular stages enable separate and scalable training on composable synthetic data, including simulated garments such as CLOTH3D and large-scale pose libraries such as AMASS for body poses and InterHand2.6M for hand poses, improving outfit generalization and pose robustness, respectively. Our approach achieves robust and expressive fitting across diverse clothing, poses, and levels of input completeness, delivering a substantial performance improvement over ETCH on both (1) seen data, such as 4D-DRESS (MPJPE-All, 33.0%) and CAPE (V2V-Hands, 35.8%), and (2) unseen data, such as BEDLAM2.0 (MPJPE-All, 80.8%; V2V-All, 80.5%).

Improvement Over ETCH

4D-DRESS MPJPE-All 33.0%
ETCH-X
ETCH
CAPE V2V-Hands 35.8%
ETCH-X
ETCH
BEDLAM2.0 MPJPE-All 80.8%
ETCH-X
ETCH
BEDLAM2.0 V2V-All 80.5%
ETCH-X
ETCH

Comparison Results with ETCH

(Scan, ETCH Result, ETCH-X Result)

Results From Partial Inputs

(Scan, Partial Input, ETCH-X Result)

Comparison Results w/o and w/ Hand Refinement

(Scan, ETCH-X w/o Hand Refinement, ETCH-X w/ Hand Refinement)

Hand refinement comparison 1
Hand refinement comparison 2
Hand refinement comparison 3

Method

ETCH Pipeline

Two stages of ETCH-X: (A) Masked Undress, (B) Dense Fit. In the Masked Undress stage, we take a clothed scan as input and compute the undressed body. In the Dense Fit stage, we implicitly learns the deforming field, which deforms the canonical SMPL-X into a posed one. Thanks to the decoupled design, the robustness to dynamic clothing and pose variations could be improved with simulated garments, i.e., CLOTH3D, and pose libraries, i.e., AMASS for body poses and InterHand2.6M for hand poses, respectively.