Existing methods introduce visible boundary artifacts due to stitching. FacEDiT produces smooth, seamless transitions.
Competing models alter unedited regions when regenerating full videos. FacEDiT preserves untouched facial motion.
All methods generate full facial motion solely from speech input.