Controllable Single-shot Animation Blending with Temporal Conditioning

Eleni Tselepi1, Spyridon Thermos2, Gerasimos Potamianos1
1ECE Dept. University of Thessaly, 2Moverse
Teaser image

We introduce a single-shot motion blending approach that uses a batched version of the GANimator, extended with a SPADE-like conditioning mechanism. Our method enables on-the-fly blending of two or more input animations by scaling (γ) and shifting (β) the motion features (M) of the skeleton convolutions, generating (G) coherent and temporally consistent motion transitions in a single forward pass, without requiring retraining or large motion datasets.

Abstract

We introduce the first single-shot motion blending model that enables seamless blending by temporally conditioning the generation process.

Our method incorporates a Spatially- Adaptive Denormalization (SPADE)-inspired conditioning scheme using skeleton-aware convolutions, enabling blending multiple input human skeletal motions into coherent animations in a single generative pass.

We evaluate our method on different datasets using both standard metrics like the Fréchet inception distance and a novel metric based on L2 error of velocity and acceleration, which we introduced to assess the smoothness of blended motion. The results demonstrate that our method produces realistic motion blending, offering a new solution to the problem of animation blending.

Video

BibTeX

@inproceedings{tselepi2025blending,
  author    = {Tselepi, Eleni and Thermos, Spyridon and Potamianos, Gerasimos},
  title     = {Controllable Single-shot Animation Blending with Temporal Conditioning},
  booktitle   = {Proceedings of the International Conference on Computer Vision  AI4VA Workshop},
  year      = {2025},
}