+
+
+
+> **[HeadArtist: Text-conditioned 3D Head Generation with Self Score Distillation, SIGGRAPH 2024](https://arxiv.org/abs/2312.07539)**
+> Hongyu Liu, Xuan Wang, Ziyu Wan, etc.
+>
+>
+>
+>
+>
+
+> **[Follow-Your-Emoji: Fine-Controllable and Expressive Freestyle Portrait Animation, SIGGRAPH Asia 2024](https://arxiv.org/abs/2406.01900)**
+> Yue Ma, Hongyu Liu, Hongfa Wang, etc.
+>
+>
+>
+
+
+ + Our approach consists of two steps during the inference process. First, the DiT model generates a 4D representation based on the input image. Then, our Motion-Aware Cross-Domain Renderer takes this 4D representation as input and, guided by both the input image and driving signals, renders it into the final target image. + +
+