Abstract
Video motion magnification techniques allow us to see small motions previously invisible to the naked eyes, such as those of vibrating airplane wings, or swaying buildings under the influence of the wind. Because the motion is small, the magnification results are prone to noise or excessive blurring. The state of the art relies on hand-designed filters to extract representations that may not be optimal. In this paper, we seek to learn the filters directly from examples using deep convolutional neural networks. To make training tractable, we carefully design a synthetic dataset that captures small motion well, and use two-frame input for training. We show that the learned filters achieve high-quality results on real videos, with less ringing artifacts and better noise characteristics than previous methods. While our model is not trained with temporal filters, we found that the temporal filters can be used with our extracted representations up to a moderate magnification, enabling a frequency-based motion selection. Finally, we analyze the learned filters and show that they behave similarly to the derivative filters used in previous works. Our code, trained model, and datasets will be available online.
T.-H. Oh and R. Jaroensri—Contributed equally.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
The ability to discern small motions enables important applications such as understanding a building’s structural health [3] and measuring a person’s vital sign [1]. Video motion magnification techniques allow us to perceive such motions. This is a difficult task, because the motions are so small that they can be indistinguishable from noise. As a result, current video magnification techniques suffer from noisy outputs and excessive blurring, especially when the magnification factor is large [24, 25, 27, 28].
Current video magnification techniques typically decompose video frames into representations that allow them to magnify motion [24, 25, 27, 28].Their decomposition typically relies on hand-designed filters, such as the complex steerable filters [6], which may not be optimal. In this paper, we seek to learn the decomposition filter directly from examples using deep convolutional neural networks (CNN). Because real motion-magnified video pairs are difficult to obtain, we designed a synthetic dataset that realistically simulates small motion. We carefully interpolate pixel values, and we explicitly model quantization, which could round away sub-level values that result from subpixel motions. These careful considerations allow us to train a network that generalizes well in real videos.
Motivated by Wadhwa et al. [24], we design a network consisting of three main parts: the spatial decomposition filters, the representation manipulator, and the reconstruction filters. To make training tractable, we simplify our training using two-frame input, and the magnified difference as the target instead of fully specifying temporal aspects of motion. Despite training on the simplified two-frames setting and synthetic data, our network achieves better noise performance and has fewer edge artifacts (See Fig. 1). Our result also suggests that the learned representations support linear operations enough to be used with linear temporal filters up to a moderate magnification factor. This enables us to select motion based on frequency bands of interest.
Finally, we visualize the learned filters and the activations to have a better understanding of what the network has learned. While the filter weights themselves show no apparent pattern, a linear approximation of our learned (non-linear) filters resembles derivative filters, which are the basis for decomposition filters in the prior art [24, 27].
The main contributions of this paper are as follows:
-
We present the first learning-based approach for the video motion magnification, which achieves high-quality magnification with fewer ringing artifacts, and has better noise characteristics.
-
We present a synthetic data generation method that captures small motions, allowing the learned filters to generalize well in real videos.
-
We analyze our model, and show that our learned filters exhibit similarity to the previously hand-engineered filters.
We will release the codes, the trained model, and the dataset online.
2 Related Work
Video Motion Magnification. Motion magnification techniques can be divided into two categories: Lagrangian and Eulerian approaches. The Lagrangian approach explicitly extracts the motion field (optical flow) and uses it to move the pixels directly [13]. The Eulerian approaches [24, 25, 27], on the other hand, decompose video frames into representations that facilitate manipulation of motions, without requiring explicit tracking. These techniques usually consist of three stages: decomposing frames into an alternative representation, manipulating the representation, and reconstructing the manipulated representation to magnified frames. Wu et al. [27] use a spatial decomposition motivated by the first-order Taylor expansion, while Wadhwa et al. [24, 25] use the complex steerable pyramid [6] to extract a phase-based representation. Current Eulerian techniques are good at revealing subtle motions, but they are hand-designed [24, 25, 27], and do not take into account many issues such as occlusion. Because of this, they are prone to noise and often suffer from excessive blurring. Our technique belongs to the Eulerian approach, but our decomposition is directly learned from examples, so it has fewer edge artifacts and better noise characteristics.
One key component of the previous motion magnification techniques is the multi-frame temporal filtering over the representations, which helps to isolate motions of interest and to prevent noise from being magnified. Wu et al. [27] and Wadhwa et al. [24, 25] utilize standard frequency bandpass filters. Their methods achieve high-quality results, but suffer from degraded quality when large motions or drifts occur in the input video. Elgharib et al. [4] and Zhang et al. [28] address this limitation. Elgharib et al. [4] model large motions using affine transformation, while Zhang et al. [28] use a different temporal processing equivalent to a second-order derivative (i.e., acceleration). On the other hand, our method achieves comparable quality even without using temporal filtering. The comparisons of our method to the prior arts are summarized in Table 1.
Deep Representation for Video Synthesis. Frame interpolation can be viewed as a complementary problem to the motion magnification problem, where the magnification factor is less than 1. Recent techniques demonstrate high-quality results by explicitly shifting pixels using either optical flow [10, 14, 26] or pixel-shifting convolution kernels [17, 18]. However, these techniques usually require re-training when changing the manipulation factor. Our representation can be directly configured for different magnification factors without re-training. For frame extrapolation, there is a line of recent work [16, 22, 23] that directly synthesizes RGB pixel values to predict dynamic video frames in the future, but their results are often blurry. Our work focusing on magnifying motion within a video, without concerns about what happens in the future.
3 Learning-Based Motion Magnification
In this section, we introduce the motion magnification problem and our learning setup. Then, we explain how we simplify the learning to make it tractable. Finally, we describe the network architecture and give the full detail of our dataset generation.
3.1 Problem Statement
We follow Wu et al. ’s and Wadhwa et al. ’s definition of motion magnification [24, 27]. Namely, given an image \(I(\mathbf {x},t) = f(\mathbf {x} +\delta (\mathbf {x}, t))\), where \(\delta (\mathbf {x}, t)\) represents the motion field as a function of position \(\mathbf {x}\) and time t, the goal of motion magnification is to magnify the motion such that the magnified image \(\tilde{I}\) becomes
where \(\alpha \) is the magnification factor. In practice, we only want to magnify certain signal \(\tilde{\delta }(\mathbf {x}, t) = \mathcal {T}(\delta (\mathbf {x}, t))\) for a selector \(\mathcal {T}(\cdot )\) that selects motion of interest, which is typically a temporal bandpass filter [24, 27].
While previous techniques rely on hand-crafted filters [24, 27], our goal is to learn a set of filters that extracts and manipulates representations of the motion signal \(\delta (\mathbf {x}, t)\) to generate output magnified frames. To simplify our training, we consider a simple two-frames input case. Specifically, we generate two input frames, \(\mathbf {X}_a\) and \(\mathbf {X}_b\) with a small motion displacement, and an output motion-magnified frame \(\mathbf {Y}\) of \(\mathbf {X}_b\) with respect to \(\mathbf {X}_a\). This reduces parameters characterizing each training pair to just the magnification factor. While this simplified setting loses the temporal aspect of motion, we will show that the network learns a linear enough representation w.r.t. the displacement to be compatible with linear temporal filters up to a moderate magnification factor.
3.2 Deep Convolutional Neural Network Architecture
Similar to Wadhwa et al. [24], our goal is to design a network that extracts a representation, which we can use to manipulate motion simply by multiplication and to reconstruct a magnified frame. Therefore, our network consists of three parts: the encoder \(G_{e}(\cdot )\), the manipulator \(G_m(\cdot )\), and the decoder \(G_d(\cdot )\), as illustrated in Fig. 2. The encoder acts as a spatial decomposition filter that extracts a shape representation [9] from a single frame, which we can use to manipulate motion (analogous to the phase of the steerable pyramid and Riesz pyramid [24, 25]). The manipulator takes this representation and manipulates it to magnify the motion (by multiplying the difference). Finally, the decoder reconstructs the modified representation into the resulting motion-magnified frames.
Our encoder and decoder are fully convolutional, which enables them to work on any resolution [15]. They use residual blocks to generate high-quality output [21]. To reduce memory footprint and increase the receptive field size, we downsample the activation by 2\(\times \) at the beginning of the encoder, and upsample it at the end of the decoder. We downsample with the strided convolution [20], and we use nearest-neighbor upsampling followed by a convolution layer to avoid checkerboard artifacts [19]. We experimentally found that three 3 \(\times \) 3 residual blocks in the encoder and nine in the decoder generally yield good results.
While Eq. (1) suggests no intensity change (constant \(f(\cdot )\)), this is not true in general. This causes our network to also magnify intensity changes. To cope with this, we introduce another output from the encoder that represents intensity information (“texture representation” [9]) similar to the amplitude of the steerable pyramid decomposition. This representation reduces undesired intensity magnification as well as noise in the final output. We downsample the representation 2\(\times \) further because it helps reduce noise. We denote the texture and shape representation outputs of the encoder as \(\mathbf {V}= G_{e,texture}(\mathbf {X})\) and \(\mathbf {M}= G_{e,shape}(\mathbf {X})\), respectively. During training, we add a regularization loss to separate these two representations, which we will discuss in more detail later.
We want to learn a shape representation \(\mathbf {M}\) that is linear with respect to \(\delta (\mathbf {x}, t)\). So, our manipulator works by taking the difference between shape representations of two given frames, and directly multiplying a magnification factor to it. That is,
In practice, we found that some non-linearity in the manipulator improves the quality of the result (See Fig. 3). Namely,
where \(g(\cdot )\) is represented by a 3 \(\times \) 3 convolution followed by ReLU, and \(h(\cdot )\) is a 3 \(\times \) 3 convolution followed by a 3 \(\times \) 3 residual block.
Loss Function. We train the whole network in an end-to-end manner. We use \(l_1\)-loss between the network output \({\hat{\mathbf {Y}}}\) and the ground-truth magnified frame \(\mathbf {Y}\). We found no noticeable difference in quality when using more advanced losses, such as the perceptual [8] or the adversarial losses [7]. In order to drive the separation of the texture and the shape representations, we perturbed the intensity of some frames, and expect the texture representations of perturbed frames to be the same, while their shape representation remain unchanged. Specifically, we create perturbed frames \(\mathbf {X}_b'\) and \(\mathbf {Y}'\), where the prime symbol indicates color perturbation. Then, we impose loses between \(\mathbf {V}_b'\) and \(\mathbf {V}_Y'\) (perturbed frames), \(\mathbf {V}_a\) and \(\mathbf {V}_b\) (un-perturbed frames), and \(\mathbf {M}_b'\) and \(\mathbf {M}_b\) (shape of perturbed frames should remain unchanged). We used \(l_1\)-loss for all regularizations. Therefore, we train the whole network G by minimizing the final loss function \(\mathcal {L}_1(\mathbf {Y},{\hat{\mathbf {Y}}}) + \lambda (\mathcal {L}_1(\mathbf {V}_a, \mathbf {V}_b) + \mathcal {L}_1(\mathbf {V}_b', \mathbf {V}_{Y}') + \mathcal {L}_1(\mathbf {M}_b, \mathbf {M}_b'))\), where \(\lambda \) is the regularization weight (set to 0.1).
Training. We use ADAM [11] with \(\beta _1 = 0.9\) and \(\beta _2 = 0.999\) to minimize the loss with the batch size 4. We set the learning rate to \(10^{-4}\) with no weight decay. In order to improve robustness to noise, we add Poisson noise with random strengths whose standard deviation is up to 3 on a \(0{-}255\) scale for a mid-gray pixel.
Applying 2-Frames Setting to Videos. Since there was no temporal concept during training, our network can be applied as long as the input has two frames. We consider two different modes where we use different frames as a reference. The Static mode uses the \(1^\mathrm {st}\) frame as an anchor, and the Dynamic uses the previous frames as a reference, i.e. we consider \((\mathbf {X}_{t-1}, \mathbf {X}_{t})\) as inputs in the Dynamic mode.
Intuitively, the Static mode follows the classical definition of motion magnification as defined in Eq. (1), while the Dynamic mode magnifies the difference (velocity) between consecutive frames. Note that the magnification factor in each case has different meanings, because we are magnifying the motion against a fixed reference, and the velocity respectively. Because there is no temporal filter, undesired motion and noise quickly becomes a problem as the magnification factor increases, and achieving high-quality result is more challenging.
Temporal Operation. Even though our network has been trained in the 2-frame setting only, we find that the shape representation is linear enough w.r.t. the displacement to be compatible with linear temporal filters. Given the shape representation \(\mathbf {M}(t)\) of a video (extracted frame-wise), we replace the difference operation with a pixel-wise temporal filter \(\mathcal {T}(\cdot )\) across the temporal axis in the manipulator \(G_{m}(\cdot )\). That is, the temporal filtering version of the manipulator, \(G_{m,temporal}(\cdot )\), is given by,
The decoder takes the temporally-filtered shape representation and the texture representation of the current frame, and generates temporally filtered motion magnified frames.
3.3 Synthetic Training Dataset
Obtaining real motion magnified video pairs is challenging. Therefore, we utilize synthetic data which can be generated in large quantity. However, simulating small motions involves several considerations because any small error will be relatively large. Our dataset is carefully designed and we will later show that the network trained on this data generalizes well to real videos. In this section, we describe considerations we make in generating our dataset.
Foreground Objects and Background Images. We utilize real image datasets for their realistic texture. We use 200, 000 images from MS COCO dataset [12] for background, and we use 7, 000 segmented objects the PASCAL VOC dataset [5] for the foreground. As the motion is magnified, filling the occluded area becomes important, so we paste our foreground objects directly onto the background to simulate occlusion effect. Each training sample contains 7 to 15 foreground objects, randomly scaled from its original size. We limit the scaling factor at 2 to avoid blurry texture. The amount and direction of motions of background and each object are also randomized to ensure that the network learns local motions.
Low Contrast Texture, Global Motion, and Static Scenes. The training examples described in the previous paragraphs are full of sharp and strong edges where the foreground and background meet. This causes the network to generalize poorly on low contrast textures. To improve generalization in these cases, we add two types of examples: where (1) the background is blurred, and (2) there is only a moving background in the scene to mimic a large object. These improve the performance on large and low contrast objects in real videos.
Small motion can be indistinguishable from noise. We find that including static scenes in the dataset helps the network learn changes that are due to noise only. We add additional two subsets where (1) the scene is completely static, and (2) the background is not moving, but the foreground is moving. With these, our dataset contains a total of 5 parts, each with 20, 000 samples of 384 \(\times \) 384 images. The examples of our dataset can be found in the supplementary material.
Input Motion and Amplification Factor. Motion magnification techniques are designed to magnify small motions at high magnifications. The task becomes even harder when the magnified motion is large (e.g. >30 pixels). To ensure the learnability of the task, we carefully parameterize each training example to make sure it is within a defined range. Specifically, we limit the magnification factor \(\alpha \) up to 100 and sample the input motion (up to 10 pixels), so that the magnified motion does not exceed 30 pixels.
Subpixel Motion Generation. How subpixel motion manifests depends on demosaicking algorithm and camera sensor pattern. Fortunately, even though our raw images are already demosaicked, they have high enough resolution that they can be downsampled to avoid artifacts from demosaicking. To ensure proper resampling, we reconstruct our image in the continuous domain before applying translation or resizing. We find that our results are not sensitive to the interpolation method used, so we chose bicubic interpolation for the reconstruction. To reduce error that results from translating by a small amount, we first generate our dataset at a higher resolution (where the motion appears larger), then downsample each frame to the desired size. We reduce aliasing when downsampling by applying a Gaussian filter whose kernel is 1 pixel in the destination domain.
Subpixel motion appears as small intensity changes that are often below the 8-bit quantization level. These changes are often rounded away especially for low contrast region. To cope with this, we add uniform quantization noise before quantizing the image. This way, each pixel has a chance of rounding up proportional to its rounding residual (e.g., if a pixel value is 102.3, it will have 30% chance of rounding up).
4 Results and Evaluations
In this section, we demonstrate the effectiveness of our proposed network and analyze its intermediate representation to shed light on what it does. We compare qualitatively and quantitatively with the state-of-the-art [24] and show that our network performs better in many aspects. Finally, we discuss limitations of our work. The comparison videos are available in our supplementary material.
4.1 Comparison with the State-of-the-Art
In this section, we compare our method with the state of the art. Because the Riesz pyramid [25] gives similar results as the steerable pyramids [24], we focus our comparison on the steerable pyramid. We perform both qualitative and quantitative evaluation as follows. All results in this section were processed with temporal filters unless otherwise noted.
Qualitative Comparison. Our method preserves edges well, and has fewer ringing artifacts. Figure 4 shows a comparison of the balance and the baby sequences, which are temporally filtered and magnified 10\(\times \) and 20\(\times \) respectively. The phase-based method shows significant ringing artifact, while ours is nearly artifact-free. This is because our representation is trained end-to-end from example motion, whereas the phase-based method relies on hand-designed multi-scale representation, which cannot handle strong edges well.
The Effect of Temporal Filters. Our method was not trained using temporal filters, so using the filters to select motion may lead to incorrect results. To test this, we consider the guitar sequence, which shows strings vibrating at different frequencies. Figure 7 shows the 25\(\times \) magnification results on the guitar sequence using different temporal filters. The strings were correctly selected by each temporal filter, which shows that the temporal filters work correctly with our representation.
Temporal processing can improve the quality of our result, because it prevents our network from magnifying unwanted motion. Figure 5 shows a comparison on the drum sequence. The temporal filter reduces blurring artifacts present when we magnify using two frames (static mode). However, even without the use of the temporal filter, our method still preserves edges well, and show no ringing artifacts. In contrast, the phase-based method shows significant ringing artifacts even when the temporal filter is applied.
Two-Frames Setting Results. Applying our network with two-frames input corresponds best to its training. We consider magnifying consecutive frames using our network (dynamic mode), and compare the result with Zhang et al. [28]. Figure 6 shows the result of gun sequence, where we apply our network in the dynamic mode without a temporal filter. As before, our result is nearly artifact free, while Zhang et al. suffers from ringing artifacts and excessive blurring, because their method is also based on the complex steerable pyramid [24]. Note that our magnification factor in the dynamic mode may have a different meaning to that of Zhang et al., but we found that for this particular sequence, using the same magnification factor (8\(\times \)) produces a magnified motion which has roughly the same size.
Quantitative Analysis. The strength of motion magnification techniques lies in its ability to visualize sub-pixel motion at high magnification factors, while being resilient to noise. To quantify these strengths and understand the limit of our method, we quantitatively evaluate our method and compare it with the phase-based method on various factors. We want to focus on comparing the representation and not temporal processing, so we generate synthetic examples whose motion is a single-frequency sinusoid and use a temporal filter that has wide passband.Footnote 1 Because our network was trained without the temporal filter, we test our method without the temporal filter, but we use temporal filters with the phase-based method. We summarize the results in Fig. 8 and its parameter ranges in the supplementary material.
For the subpixel motion test, we generate synthetic data having foreground input motion ranging from 0.01 to 1 pixel. We vary the magnification factor \(\alpha \) such that the magnified motion is 10 pixels. No noise was added. Additionally, we move the background for the same amount of motion but in a different direction to all foreground objects. This ensures that no method could do well by simply replicating the background.
In the noise test, we fixed the amount of input motion and magnification factor and added noise to the input frames. We do not move background in this case. To simulate photon noise, we create a noise image whose variance equals the value of each pixel in the original image. A multiplicative noise factor controls the final strength of noise image to be added.
Because the magnified motion is not very large (10 pixels), the input and the output magnified frames could be largely similar. We also calculate the SSIM between the input and output frames as a baseline reference in addition to the phase-based method.
In all tests, our method performs better than the phase-based method. As Fig. 8(a) shows, our sub-pixel performance remains high all the way down to 0.01 pixels, and it exceeds 1 standard deviation of the phase-based performance as the motion increase above 0.02 pixels. Interestingly, despite being trained only up to 100\(\times \) magnification, the network performs considerably well at the smallest input motion (0.01), where magnification factor reaches 1,000\(\times \). This suggests that our network are more limited by the amount of output motion it needs to generate, rather than the magnification factors it was given.
Figure 8(b, c) show the test results under noisy conditions with different amounts of input motion. In all cases, the performance of our method is consistently higher than that of the phase-based method, which quickly drops to the level of the baseline as the noise factor increase above 1.0. Comparing across different input motion, our performance degrades faster as the input motion becomes smaller (See Fig. 8(b, c)). This is expected because when the motion is small, it becomes harder to distinguish actual motion from noise. Some video outputs from these tests are included in the supplementary material.
4.2 Physical Accuracy of Our Method
In nearly all of our real test videos, the resulting motions produced by our method have similar magnitude as, and are in phase with, the motions produced by [24] (see Fig. 1, and the supplementary videos). This shows that our method is at least as physically accurate as the phase-based method, while exhibiting fewer artifacts.
We also obtained the hammer sequence from the authors of [24], where accelerometer measurement was available. We integrated twice the accelerometer signal and used a zero-phase high-pass filter to remove drifts. As Fig. 10 shows, the resulting signal (blue line) matches up well with our 10\(\times \) magnified (without temporal filter) result, suggesting that our method is physically accurate.
4.3 Visualizing Network Activation
Deep neural networks achieve high performance in a wide variety of vision tasks, but their inner working is still largely unknown [2]. In this section, we analyze our network to understand what it does, and show that it extracts relevant information to the task. We analyze the response of the encoder, by approximating it as a linear system. We pass several test images through the encoder, and calculate the average impulse responses across the images. Figure 9 shows the samples of the linear kernel approximation of the encoder’s shape response. Many of these responses resemble Gabor filters and Laplacian filters, which suggests that our network learns to extract similar information as done by the complex steerable filters [24]. By contrast, the texture kernel responses show many blurring kernels.
4.4 Limitations
While our network performs well in the 2-frame setting, its performance degrades with temporal filters when the magnification factor is high and motion is small. Figure 11 shows an example frame of temporally-filtered magnified synthetic videos with increasing the magnification factor. As the magnification factor increases, blurring becomes prominent, and strong color artifacts appear as the magnification factor exceeds what the network was trained on.
In some real videos, our method with temporal filter appears to be blind to very small motions. This results in patchy magnification where some patches get occasionally magnified when their motions are large enough for the network to see. Figure 12 shows our magnification results of the eye sequence compared to that of the phase-based method [24]. Our magnification result shows little motion, except on a few occasions, while the phase-based method reveals a richer motion of the iris. We expect to see some artifact on our network running with temporal filters, because it was not what it was trained on. However, this limits its usefulness in cases where the temporal filter is essential to selecting small motion of interest. Improving compatibility with the temporal filter will be an important direction for future work.
5 Conclusion
Current motion magnification techniques are based on hand-designed filters, and are prone to noise and excessive blurring. We present a new learning-based motion magnification method that seeks to learn the filters directly from data. We simplify training by using the two-frames input setting to make it tractable. We generate a set of carefully designed synthetic data that captures aspects of small motion well. Despite these simplifications, we show that our network performs well, and has less edge artifact and better noise characteristics than the state of the arts. Our method is compatible with temporal filters, and yielded good results up to a moderate magnification factor. Improving compatibility with temporal filters so that it works at higher magnification is an important direction for future work.
Notes
- 1.
Our motion is 3 Hz at 30 fps, and the temporal filter used is a 30-tap FIR with a passband between 0.5–7.5 Hz.
References
Balakrishnan, G., Durand, F., Guttag, J.: Detecting pulse from head motions in video. In: IEEE Conference on Computer Vision and Pattern Recognition (2013)
Bau, D., Zhou, B., Khosla, A., Oliva, A., Torralba, A.: Network dissection: quantifying interpretability of deep visual representations. In: IEEE Conference on Computer Vision and Pattern Recognition (2017)
Cha, Y.J., Chen, J., Büyüköztürk, O.: Output-only computer vision based damage detection using phase-based optical flow and unscented Kalman filters. Eng. Struct. 132, 300–313 (2017)
Elgharib, M.A., Hefeeda, M., Durand, F., Freeman, W.T.: Video magnification in presence of large motions. In: IEEE Conference on Computer Vision and Pattern Recognition (2015)
Everingham, M., Van Gool, L., Williams, C.K.I., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 88(2), 303–338 (2010)
Freeman, W.T., Adelson, E.H.: The design and use of steerable filters. IEEE Trans. Pattern Anal. Mach. Intell. 13(9), 891–906 (1991)
Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: IEEE Conference on Computer Vision and Pattern Recognition (2017)
Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694–711. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_43
Jones, M.J., Poggio, T.: Multidimensional morphable models: a framework for representing and matching object classes. Int. J. Comput. Vis. 29(2), 107–131 (1998)
Kalantari, N.K., Wang, T.C., Ramamoorthi, R.: Learning-based view synthesis for light field cameras. ACM Trans. Graph. (SIGGRAPH Asia) 35(6), 193–10 (2016)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740–755. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10602-1_48
Liu, C., Torralba, A., Freeman, W.T., Durand, F., Adelson, E.H.: Motion magnification. ACM Trans. Graph. (SIGGRAPH) 24(3), 519–526 (2005)
Liu, Z., Yeh, R.A., Tang, X., Liu, Y., Agarwala, A.: Video frame synthesis using deep voxel flow. In: IEEE International Conference on Computer Vision (2017)
Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition (2015)
Mathieu, M., Couprie, C., LeCun, Y.: Deep multi-scale video prediction beyond mean square error. In: International Conference on Learning Representations (2016)
Niklaus, S., Mai, L., Liu, F.: Video frame interpolation via adaptive convolution. In: IEEE Conference on Computer Vision and Pattern Recognition (2017)
Niklaus, S., Mai, L., Liu, F.: Video frame interpolation via adaptive separable convolution. In: IEEE International Conference on Computer Vision (2017)
Odena, A., Dumoulin, V., Olah, C.: Deconvolution and checkerboard artifacts. Distill 1(10), e3 (2016)
Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)
Sajjadi, M.S., Schölkopf, B., Hirsch, M.: EnhanceNet: single image super-resolution through automated texture synthesis. In: IEEE International Conference on Computer Vision (2017)
Srivastava, N., Mansimov, E., Salakhudinov, R.: Unsupervised learning of video representations using LSTMs. In: International Conference on Machine Learning (2015)
Villegas, R., Yang, J., Hong, S., Lin, X., Lee, H.: Decomposing motion and content for natural video sequence prediction. In: International Conference on Learning Representations (2017)
Wadhwa, N., Rubinstein, M., Durand, F., Freeman, W.T.: Phase-based video motion processing. ACM Trans. Graph. (SIGGRAPH) 32(4), 80 (2013)
Wadhwa, N., Rubinstein, M., Durand, F., Freeman, W.T.: Riesz pyramids for fast phase-based video magnification. In: IEEE International Conference on Computational Photography (2014)
Wang, T., Zhu, J., Kalantari, N.K., Efros, A.A., Ramamoorthi, R.: Light field video capture using a learning-based hybrid imaging system. ACM Trans. Graph. (SIGGRAPH) 36(4), 133:1–133:13 (2017)
Wu, H.Y., Rubinstein, M., Shih, E., Guttag, J., Durand, F., Freeman, W.: Eulerian video magnification for revealing subtle changes in the world. ACM Trans. Graph. (SIGGRAPH) 31(4), 65–8 (2012)
Zhang, Y., Pintea, S.L., van Gemert, J.C.: Video acceleration magnification. In: IEEE Conference on Computer Vision and Pattern Recognition (2017)
Acknowledgment
The authors would like to thank Qatar Computing Research Institute, Toyota Research Institute, and Shell Research for their generous support of this project. Changil Kim was supported by a Swiss National Science Foundation fellowship P2EZP2 168785.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Oh, TH. et al. (2018). Learning-Based Video Motion Magnification. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds) Computer Vision – ECCV 2018. ECCV 2018. Lecture Notes in Computer Science(), vol 11208. Springer, Cham. https://doi.org/10.1007/978-3-030-01225-0_39
Download citation
DOI: https://doi.org/10.1007/978-3-030-01225-0_39
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-01224-3
Online ISBN: 978-3-030-01225-0
eBook Packages: Computer ScienceComputer Science (R0)