HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

FlowFormer: A Transformer Architecture for Optical Flow

Zhaoyang Huang Xiaoyu Shi Chao Zhang Qiang Wang Ka Chun Cheung Hongwei Qin Jifeng Dai Hongsheng Li

FlowFormer: A Transformer Architecture for Optical Flow

Abstract

We introduce optical Flow transFormer, dubbed as FlowFormer, a transformer-based neural network architecture for learning optical flow. FlowFormer tokenizes the 4D cost volume built from an image pair, encodes the cost tokens into a cost memory with alternate-group transformer (AGT) layers in a novel latent space, and decodes the cost memory via a recurrent transformer decoder with dynamic positional cost queries. On the Sintel benchmark, FlowFormer achieves 1.159 and 2.088 average end-point-error (AEPE) on the clean and final pass, a 16.5% and 15.5% error reduction from the best published result (1.388 and 2.47). Besides, FlowFormer also achieves strong generalization performance. Without being trained on Sintel, FlowFormer achieves 1.01 AEPE on the clean pass of Sintel training set, outperforming the best published result (1.29) by 21.7%.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
optical-flow-estimation-on-kitti-2015-trainFlowFormer
EPE: 4.09
F1-all: 14.7
optical-flow-estimation-on-sintel-cleanFlowFormer
Average End-Point Error: 1.16
optical-flow-estimation-on-springFlowFormer
1px total: 6.510

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp