Learning-based Video Motion Magnification

April 08, 2018 ยท Entered Twilight ยท ๐Ÿ› European Conference on Computer Vision

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"No code URL or promise found in abstract"
"Code repo scraped from project page (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, LICENSE, README.md, configs, convert_3frames_data_to_tfrecords.py, data_loader.py, magnet.py, main.py, modules.py, ops.py, preprocessor.py, requirements.txt, run_on_test_videos.sh, run_temporal_on_test_videos.sh, train.sh, utils.py

Authors Tae-Hyun Oh, Ronnachai Jaroensri, Changil Kim, Mohamed Elgharib, Frรฉdo Durand, William T. Freeman, Wojciech Matusik arXiv ID 1804.02684 Category cs.CV: Computer Vision Cross-listed cs.GR Citations 181 Venue European Conference on Computer Vision Repository https://github.com/12dmodel/deep_motion_mag โญ 498 Last Checked 1 month ago
Abstract
Video motion magnification techniques allow us to see small motions previously invisible to the naked eyes, such as those of vibrating airplane wings, or swaying buildings under the influence of the wind. Because the motion is small, the magnification results are prone to noise or excessive blurring. The state of the art relies on hand-designed filters to extract representations that may not be optimal. In this paper, we seek to learn the filters directly from examples using deep convolutional neural networks. To make training tractable, we carefully design a synthetic dataset that captures small motion well, and use two-frame input for training. We show that the learned filters achieve high-quality results on real videos, with less ringing artifacts and better noise characteristics than previous methods. While our model is not trained with temporal filters, we found that the temporal filters can be used with our extracted representations up to a moderate magnification, enabling a frequency-based motion selection. Finally, we analyze the learned filters and show that they behave similarly to the derivative filters used in previous works. Our code, trained model, and datasets will be available online.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision