Demajh logo Demajh, Inc.

GeoAvatar: Adaptive Geometrical Gaussian Splatting for 3D Head Avatar: what it means for business leaders

GeoAvatar transforms ordinary video into lifelike 3-D head avatars by letting tiny Gaussian point-clouds flex where human faces vary most, while locking down areas that should stay rigid — delivering realism without ballooning compute.

1. What the method is

GeoAvatar is a pipeline that converts a brief monocular video into a textured, animatable 3-D head. Instead of a fixed mesh, it deploys thousands of translucent Gaussians whose positions, sizes, colours and opacities are jointly optimised. A preprocessing step tags each Gaussian as rigid or flexible: rigid splats hug bone-anchored regions, while flexible ones track hair, ears and expressive skin. A dedicated mouth rig injects palate, molars and jaw floor geometry for accurate speech animation. The resulting model renders in real time on consumer GPUs thanks to its compact parameter footprint.

2. Why the method was developed

Existing avatar systems juggle a trade-off: rigid meshes keep identity but look plastic, whereas free Gaussians capture detail yet wobble when animated. The authors realised deviation tolerance should depend on local geometric certainty, so they designed an adaptive constraint scheme—tight where morphable models are trustworthy, loose where they are not. This suppresses lip-sync drift and hair popping, trimming post-production time for virtual presenters, digital humans and AR try-ons.

3. Who should care

Studios creating virtual influencers, XR platforms demanding low-latency head rendering, enterprise teams prototyping customer-service avatars and game producers seeking actor-specific facial capture all stand to gain. Chip vendors targeting on-device avatar calls will also appreciate GeoAvatar’s small model size and rapid inference.

4. How the method works

The pipeline first fits a FLAME morphable head to each frame for coarse geometry. Gaussians are then scattered across mesh faces and evaluated against pixel reprojection error. Low-error splats enter the rigid set and receive a tight positional penalty; high-error splats join the flexible pool with softer regularisation. Mouth parts are grouped into upper and lower units that share deformation vectors, ensuring anatomical coherence as the jaw moves. End-to-end optimisation employs differentiable rasterisation, photometric loss and a rigging term that tethers each Gaussian to its nearest mesh triangle.

5. How it was evaluated

Tests ran on three public monocular-video sets plus the new DynamicFace corpus spanning extreme expressions and head turns. Metrics included reconstruction PSNR, identity similarity and animation quality. Baselines were SplattingAvatar, FlashAvatar and other Gaussian-splat methods, all trained on a single RTX 4090 to reflect indie-studio constraints.

6. How it performed

GeoAvatar cut animation artifacts by 38 % and raised mouth-region PSNR by 2–5 dB versus the best baseline, all while keeping inference below 15 ms per frame. Rigid–flex partitioning preserved hair realism under extreme poses, and the mouth rig eliminated interior tearing seen in prior work. The complete avatar weighs under 25 MB, fitting mobile deployments. (Source: arXiv 2507.18155, 2025)

← Back to dossier index