AGORA adopts a 3D GAN framework to learn an animatable 3D head generation model from 2D image datasets. The architecture consists of two main components:
The final 3D position of each Gaussian is obtained via 3D lifting: we interpolate a base position from the articulated FLAME mesh at UV coordinates and add the predicted offset. This design anchors the generated 3DGS to the underlying parametric mesh, providing a structured basis for animation.
To enforce expression consistency, we adopt a dual-discrimination scheme. The discriminator is conditioned on the target expression by concatenating the rendered image with a synthetic rendering of the FLAME mesh, where vertices are color-coded by their expression-isolated vertex displacement from the neutral pose. This allows the discriminator to penalize fine-grained deviations in expression.
Generated avatars (seeds 0-32) reenacted by the driving video on the left.
Avatars generated from single images, driven by the video on the left.
@article{fazylov2025agora,
author = {Fazylov, Ramazan and Zagoruyko, Sergey and Parkin, Aleksandr and Lefkimmiatis, Stamatis and Laptev, Ivan},
title = {{AGORA: Adversarial Generation Of Real-time Animatable 3D Gaussian Head Avatars}},
journal = {arXiv preprint arXiv: NOT YET ON ARXIV, ETA FEW DAYS},
year = {2025}
}