Full-Frame Video Stabilization via Spatiotemporal Transformers


Karacan L., Sarıgül M.

Computational Visual Media, cilt.11, sa.3, ss.655-667, 2025 (SCI-Expanded)

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 11 Sayı: 3
  • Basım Tarihi: 2025
  • Doi Numarası: 10.26599/cvm.2025.9450416
  • Dergi Adı: Computational Visual Media
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED)
  • Sayfa Sayıları: ss.655-667
  • Çukurova Üniversitesi Adresli: Evet

Özet

Traditional video stabilization methods use a warping operation to smooth the camera path but result in missing regions in the video frames. To solve this issue, full-frame video stabilization techniques attempt to fill in the unidentified boundary regions, but their effectiveness is limited. In this work, we propose a full-frame video stabilization method using spatiotemporal transformers to fill the missing boundary regions after the warping operation. For training, we adopt a self-supervised strategy and improve it by incorporating temporal information. The proposed approach allows the utilization of redundant video information spatially and temporally while filling in missing regions. Experimental results show that our approach achieves superior results on popular video stabilization datasets. The code, pre-trained model, and video results are available at https://github.com/leventkaracan/VidStabFormer.