Universal Prototype Transport for Zero-Shot Action Recognition and Localization

This work addresses the problem of recognizing action categories in videos when no training examples are available. The current state-of-the-art enables such a zero-shot recognition by learning universal mappings from videos to a semantic space, either trained on large-scale seen actions or on objects. While effective, we find that universal action and object mappings are biased to specific regions in the semantic space. These biases lead to a fundamental problem: many unseen action categories are simply never inferred during testing. For example on UCF-101, a quarter of the unseen actions are out of reach with a state-of-the-art universal action model. To that end, this paper introduces universal prototype transport for zero-shot action recognition. The main idea is to re-position the semantic prototypes of unseen actions by matching them to the distribution of all test videos. For universal action models, we propose to match distributions through a hyperspherical optimal transport from unseen action prototypes to the set of all projected test videos. The resulting transport couplings in turn determine the target prototype for each unseen action. Rather than directly using the target prototype as final result, we re-position unseen action prototypes along the geodesic spanned by the original and target prototypes as a form of semantic regularization. For universal object models, we outline a variant that defines target prototypes based on an optimal transport between unseen action prototypes and object prototypes. Empirically, we show that universal prototype transport diminishes the biased selection of unseen action prototypes and boosts both universal action and object models for zero-shot classification and spatio-temporal localization.

Abstract This work addresses the problem of recognizing action categories in videos when no training examples are available.The current state-of-the-art enables such a zero-shot recognition by learning universal mappings from videos to a semantic space, either trained on large-scale seen actions or on objects.While effective, we find that universal action and object mappings are biased to specific regions in the semantic space.These biases lead to a fundamental problem: many unseen action categories are simply never inferred during testing.For example on UCF-101, a quarter of the unseen actions are out of reach with a state-of-the-art universal action model.To that end, this paper introduces universal prototype transport for zero-shot action recognition.The main idea is to re-position the semantic prototypes of unseen actions by matching them to the distribution of all test videos.For universal action models, we propose to match distributions through a hyperspherical optimal transport from unseen action prototypes to the set of all projected test videos.The resulting transport couplings in turn determine the target prototype for each unseen action.Rather than directly using the target prototype as final result, we re-position unseen action prototypes along the geodesic spanned by the original and target prototypes as a form of semantic regularization.For universal object models, we outline a variant that defines target prototypes based on an optimal transport between unseen action prototypes and object prototypes.Empirically, we show that universal prototype transport diminishes the biased selection of unseen action prototypes and boosts both universal ac-

Introduction
This paper addresses the problem of recognizing actions in videos.Foundational deep network approaches performed action recognition through frame-level fusion (Karpathy et al, 2014), two-stream networks (Simonyan and Zisserman, 2014;Feichtenhofer et al, 2016), and 3D convolutional networks (Carreira and Zisserman, 2017).Building upon these approaches, recent works have shown great recognition capabilities through e.g., slow-fast architectures (Feichtenhofer et al, 2019), separated 3D convolutions (Tran et al, 2019), and video transformers (Arnab et al, 2021).Such deep networks require large amounts of video material for training and efforts have been made to meet those video demands, such as ActivityNet (Caba Heilbron et al, 2015), EPIC Kitchens (Damen et al, 2018), Kinetics (Carreira and Zisserman, 2017), HowTo100M (Miech et al, 2019), and EGO4D (Grauman et al, 2022) to name a few.While such datasets increase the coverage of the action space, we seek to recognize actions even when no examples are available during training.
In zero-shot action recognition, many works have outlined approaches that mirror successes in the image domain, for example by using attributes (Liu et al, 2011;Gan et al, 2016b) or feature synthesis (Mishra et al, 2020) to transfer knowledge from seen to unseen actions.More recently, state-of-the-art results have been achieved by taking a universal learning perspective, where large-scale models are trained to map input videos to a shared semantic space occupied by both seen and unseen categories.In the first universal perspec-

Ground truth
Fig. 1: Predicted and ground truth action distributions for an example universal action model.The distributions are out of alignment, with a quarter of the actions never predicted altogether, severely limiting the upper bound for zero-shot action recognition.We seek to overcome this misalignment by matching the distributions of unseen actions and test videos.
tive, large-scale networks train a mapping from videos to a semantic space on hundreds of seen actions from e.g., ActivityNet (Zhu et al, 2018) or Kinetics (Brattoli et al, 2020;Pu et al, 2022).For a target dataset with unseen actions, zero-shot inference is directly possible through a nearest neighbour search between the video mappings and the embeddings of unseen actions.In the second perspective, networks are trained on thousands of objects (Jain et al, 2015;Mettes et al, 2021), where inference becomes a function of the object likelihoods in test videos and the semantic relation between objects and actions.
While effective, both universal perspectives share a common limitation: they are strongly biased to subsets of the unseen actions.Compounding biases in (i) the mapping of videos to seen categories and (ii) in the matching between seen and unseen categories result in a mismatch between the prototypes of unseen actions and projected test videos.A significant part of the actions are as a result simply never selected, disrupting progress in zero-shot action recognition.This is a direct result of their inductive nature, where each test video is individually evaluated.Figure 1 illustrates this problem on UCF-101 for example with the recent universal action model of Brattoli et al (2020).Here, 23% (!) of the unseen actions are simply never selected for any test video in the first place, placing hard upper bounds on the accuracy in zero-shot recognition.We seek to ad-dress this limitation by enriching universal models with an optimal transport transductive perspective.
We introduce universal prototype transport for zeroshot action recognition with universal models.The main idea is to re-position the prototypes of unseen actions in the shared semantic space to better align with the test videos.For universal action models, we first find an optimal mapping from action prototypes to test videos on the hypersphere.We then define a target prototype for each unseen action as a weighted Fréchet mean based on the coupling matrix from the optimal transport.Rather than doing a nearest neighbor search directly on the target prototypes for zero-shot recognition, we re-position the prototypes through interpolation along the geodesic spanned by the original and target prototypes.The intuition behind the interpolation is to maintain a form of semantic regularization for the target prototypes from their original semantic representation.For universal object models, we follow a similar setup, but replace the test videos with object prototypes during the optimal transport mapping.Beyond action recognition, we show that our approach is also helpful for zero-shot spatiotemporal action localization due to an improved ranking of tubes from different test videos to unseen actions.
We perform empirical evaluations on four action datasets for two tasks: zero-shot action recognition and zero-shot spatio-temporal action localization.The experiments confirm that universal prototype transport diminishes the biased selection of unseen actions in universal models, resulting in better recognition performance.By combining inductive universal action and object models, we are able to improve both zero-shot action recognition and spatio-temporal action localization in videos.Our approach is general in nature and can be used on top of any existing approach.The code will be made available on github.

Zero-shot action recognition
Zero-shot action recognition refers to the task of assigning an action label to a test video given a pool of actions not observed during training.A common approach is to learn and transfer a shared representation on seen actions with training examples to be able to perform inference on unseen actions.A well-known shared space is given by attributes (Gan et al, 2016b;Liu et al, 2011;Zhang et al, 2015).By projecting test videos to an attribute space, inference is possible through a neighbour search with unseen actions manually defined in the same space.Since attributes require manual annotations for every action, other works prefer to use word embeddings (Bishay et al, 2019;Gan et al, 2016a,c;Li et al, 2016), video captions (Estevam et al, 2021a), or action hierarchies (Long et al, 2020) to provide a shared space for seen and unseen actions.
State-of-the-art zero-shot action recognition solutions take a universal learning perspective with semantic word embeddings as the shared space for knowledge transfer.Rather than relying on a small set of seen actions from the same dataset, large-scale models are trained on hundreds or thousands of seen categories to learn a direct mapping from videos to the shared space.The first universal learning direction relies on large-scale actions with training videos.Zhu et al (2018) were the first to propose a large-scale universal action perspective by learning a video network on 200 actions from ActivityNet (Caba Heilbron et al, 2015).Brattoli et al (2020) have obtained high performance in zero-shot action recognition by scaling this perspective to 664 actions from Kinetics (Carreira and Zisserman, 2017).Pu et al (2022) have subsequently shown that incorporating alignment, uniformity, and feature synthesis further improve recognition.Due to the large amount of seen actions, care needs to be taken to avoid (near) duplicates between seen and unseen actions, as also noted by Roitberg et al (2018).These universal models share a similar fate, where the predicted and ground truth action distributions are mis-aligned due to biased training.We show that our proposed prototype transport improves the current state-of-the-art actionbased models for zero-shot recognition.
Next to universal action models, several works have shown competitive results by taking a universal object perspective.Large-scale networks are trained in the image domain on thousands of object labels (Jain et al, 2015;Liu et al, 2019;Mettes and Snoek, 2017;Mettes et al, 2021;Wu et al, 2016).Once trained any action can be inferred based on the object likelihoods in test video and the semantic similarities between objects and unseen actions.For example, Bretti and Mettes (2021) show that actions like skateboarding are easy to recognize as its corresponding action embedding is close to object embeddings like skateboard and roller skates.Similar to universal action models, we show that the object-based perspective also benefits from our proposed approach.
By operating on the entire test distribution, we take a transductive view, a common setting in zero-shot action recognition as noted by Estevam et al (2021b).Where inductive reasoning requires solving a general problem and applying them to individual samples, transduction is about reasoning from observed training cases to observed test cases.It is hence commonly seen as a more direct way to solve inference problems (Vapnik, 2006) with direct implications for zero-shot learning, semi-supervised learning, transfer learning, and more (Arnold et al, 2007).In practical settings, inference is often performed on video collections, for example for recommendation or searching in large databases, making transductive learning a viable learning setting.Rohrbach et al (2013) provide a foundation for transduction in zero-shot context by exploiting the intersample similarity over the test set.In similar spirit, several works have proposed transductive extensions for zero-shot action recognition (Alexiou et al, 2016;Fu et al, 2014;Gao et al, 2019;Kodirov et al, 2015;Mandal et al, 2019;Xu et al, 2017Xu et al, , 2020;;Zhuo et al, 2022).Transductive learning performs inference over the entire video batch, rather than each video individually.Different from existing approaches, we use the test video distribution to improve the unseen action embeddings in the shared space of universal models by building on optimal transport.Moreover, our approach can switch between inductive and transductive settings, as only the positions of unseen action prototypes are updated.
Wang et al ( 2021) have also investigated optimal transport in the context of zero-shot learning.In their work, optimal transport is used to match distributions of generated and real features in their generative learning.Wu et al (20212) consider optimal transport to align batches of unpaired images and textual descriptions for self-supervised learning with zero-shot learning as down-stream task.In our work, we outline a hyperspherical optimal transport to align distributions of unseen action labels to videos and objects in shared semantic spaces for zero-shot recognition in videos.

Zero-shot action localization
Beyond recognition, a number of works have researched action localization in zero-shot settings.In the temporal domain, zero-shot localization has been investigated by aligning temporal proposals with label embeddings (Zhang et al, 2020(Zhang et al, , 2022)), by training models on trimmed seen actions followed by a knowledge transfer to unseen actions (Jain et al, 2020), or by taking an open-set perspective (Bao et al, 2022).
In the spatio-temporal domain, Jain et al (2015) were the first to investigate zero-shot spatio-temporal action localization with a universal object perspective by computing object likelihoods for spatio-temporal proposals, followed by a semantic transfer to unseen actions.This direction has been expanded by incorporating knowledge about spatial relations between actors and objects (Mettes and Snoek, 2017) and by taking into account semantic priors about objects (Mettes et al, 2021).We build upon these works and improve the ranking of action tubes from different videos through our universal prototype transport.Soomro and Shah (2017) have previously performed unsupervised spatio-temporal action localization, while Yang et al (2022) have recently enabled a similar localization by performing a self-shot learning to automatically find relevant common videos from an unlabelled video pool to help the optimization.Different from these works, we seek to perform zero-shot spatiotemporal localization by assigning semantic labels to action tubes, rather than unsupervised or common action localization.

Background
This paper builds upon optimal transport to improve zero-shot recognition.The background section provides the background for the general problem and the common task to be solved in optimal transport, following the formulation of Peyré et al (2019).Optimal transport is the problem of moving one distribution of mass to another with minimal effort, with piles of dirt and multiple holes as a classical practical example.
More formally, optimal transport is a minimization problem over discrete measures, where a discrete measure is defined as: with δ xi the Dirac position of the i th element and a i ∈ Σ n denotes probability vector and element of the probability simplex: Optimal transport is concerned with finding an optimal assignment between two discrete measures.If we assume that two discrete measures are of equal size and want to find a one-to-one mapping between the elements of the two measures, we arrive at the Monge problem (Monge, 1781): with C i,j a precomputed cost matrix between two discrete measures and Perm(•) the set of all possible permuations.Here, we are interested in optimal assignment between discrete measures of different sizes and distributing mass from any point in one discrete measure to multiple points in the other discrete measure.
We will therefore focus on the Kantorovich relaxation of the Monge problem (Kantorovich, 1942).In this relaxation, the permutation operation is replaced by a coupling matrix P ∈ R n×m

+
, where P i,j denotes the amount of mass that is distributed from point i to point j.The minimization problem is in turn given as: with a 1 and a 2 two discrete measures and U(a 1 , a 2 ) the set of possible admissible couplings.With a 1 , a 2 , and C known, the goal is to find the optimal coupling matrix.We will extensively rely on optimal transport on the hypersphere and on the coupling matrix of the Kantorovich relaxation in the context of zero-shot actions.For a full foundation on optimal transport, we recommend the work of Peyré et al (2019).

Universal prototype transport
For the problem of zero-shot action recognition, we are given a set of test videos V u and a set of labels L u denoting actions which have not been observed during training.We seek to assign a label l ∈ L u to each test video.We start from two state-of-the-art universal learning directions in zero-shot action recognition, namely by transferring knowledge from large-scale seen actions and from objects.Below, we first introduce transductive universal transport for the transfer from seen to unseen actions.Second, we extend our approach for re-positioning unseen action embeddings based on universal object models.

Transporting universal action models
Universal action models are centered around a semantic space that should be shared by both videos and action labels.This requires two transformation functions: a function ω that maps a label to a prototype in the semantic embedding space and a function ϕ that maps a video to the same embedding space.The function ω is given by a pre-trained word embedding model (Mikolov et al, 2013b), where embeddings are ℓ 2 -normalized and optimized with the cosine distance.The mapping function ϕ is learned on a set of training videos V s with seen action labels L s , with its loss given as: with l v ∈ L s the action label for video v. Once such a network is optimized for a set of seen actions with training videos, zero-shot learning can be enabled for Fig. 2: Overview of universal prototype transport.First, we find an optimal mapping from unseen action prototypes to the projected test videos when building on universal action models.For universal object models, the test videos are replaced by object prototypes.Second, we define the target prototype for each unseen action as the weighted Fréchet mean over the transport couplings.Third, we re-position unseen action prototypes along the geodesic spanned by the original and target prototypes.
a test video simply through a nearest neighbor search with a set of unseen action prototypes in the shared semantic space.This paper argues that that since ϕ is trained on actions not used during inference, the projected videos and unseen action prototypes are not well aligned.We propose to improve zero-shot action recognition by re-positioning unseen action prototypes with optimal transport.The main idea here is to find an optimal mapping between the set of unseen action prototypes and the set of projected test videos in the shared semantic space.We then want to utilize the optimal mapping to improve the location of unseen action prototypes in the shared space, to improve the zero-shot nearest neighbor inference.Figure 2 provides an overview of our approach.To be able to find the optimal mapping between unseen actions and test videos, we first need to redefine them as discrete measures in order to solve the corresponding optimal transport problem.For the unseen actions, the definition is given as: Definition 1 (Actions as a discrete measure).The set of unseen actions is represented as a measure as: where w lu ∈ w u denotes the set of weights for the action, such that w u ∈ Σ |Lu| is on the probability simplex, and l u denotes the label of unseen action u.
The unseen actions are given by a weighted combination of their word embeddings in the shared semantic space.The definition of the projected test videos is given as: Definition 2 (Videos as a discrete measure).The set of projected test videos is given as a measure as: with w c ∈ w c ∈ Σ k akin to Definition 1, where C k denotes a k-component cluster aggregation over the set of videos, and where a(V u ; c) denotes the set of videos assigned to cluster c.
Rather than using all test videos as individual points in the discrete measure, we first cluster the test videos and define the measure over the cluster centers.The idea behind Definition 2 is to make the discrete measure robust to outliers and to increase the focus on highdensity regions of projected videos.
With the unseen action labels and videos defined as discrete measures in the same space, we are able to compute an optimal transport mapping between the two.By operating over the entire distribution of test videos rather than performing inference for each video independently, we view this as a transductive form of optimal transport.We seek to obtain a coupling matrix P u with the minimization objection given in Equation 4. To that end, let C u denote the required cost matrix, with C u ij defined as the cosine distance between unseen action i and video cluster j.Then the hyperspherical optimal transport from µ u to µ v is as: with P u ij a single coupling value, where the minimization is solved using the Lagrangian approach of Bon-neel et al (2011).This results in a coupling matrix P u .When working with universal action models, we set the weights w u and w v uniformly.The overall step of finding an optimal mapping from unseen actions to test videos is shown in Figure 2 as step (1) in the universal prototype transport.
Given the optimal transport mapping, we propose to condense the corresponding coupling into a single target prototype per unseen action.Since the semantic space on which we operate is a hypersphere, we define the target prototype for unseen action i as the weighted Fréchet mean (Lou et al, 2020;Miolane et al, 2020) based on normalized coupling values: with d the cosine similarity and s the obtained mean.Determining the target prototype of each unseen action is visualized in Figure 2 as step (2).We opt for a hyperspherical optimal transport formulation because we rely on word embeddings for actions and objects, which are hyperspherical in nature as they are optimized with cosine distances (Mikolov et al, 2013b) and they are state-of-the-art for zero-shot action recognition (Brattoli et al, 2020;Pu et al, 2022;Zhuo et al, 2022).The target provides a new prototype in embedding space for each unseen action, guided by the distribution of mapped and clustered test videos.While we can directly use the new embeddings for inference, we want to avoid big changes in embedding space since that relates with losing the original semantic interpretation of the action.We therefore dictate that the final prototype of each unseen action is positioned along the geodesic spanned by the original and target prototypes, modelled through spherical interpolation: where λ denotes the interpolation ratio between the extremes.In this manner, we move each unseen action towards its proposed target prototype, with the interpolation acting as a regularization pulling the action towards the original semantic prototype, visualized as step (3) in Figure 2. Zero-shot inference is performed the same as in existing universal action models, by means of a nearest neighbour search between video v and each unseen action label l as after which the action label with the highest similarity is selected.

Transporting universal object models
Universal object models for zero-shot action recognition suffer from the same bias in the assignment of unseen action labels to test videos.We therefore also seek to transport unseen action prototypes in this setting.We start by redefining objects as discrete measures to make them suitable in the context of optimal transport: Definition 3 (Objects as a discrete measure).The set of objects are given as a measure as : with w o ∈ o c and where p(o|v) denotes the likelihood of object o occurring in video v.
Unique in this definition, the discrete measure for objects is based on a subset O s ∈ O, i.e., we define a degenerate distribution over objects.The subset is determined by again taking a transductive view; we exclude any object which does not have a likelihood over a low threshold τ in any test video.The idea behind this is simple: we want to avoid a bias in the optimal transport to objects which do not actually occur in videos.
In the universal object context, the optimal transport mapping is now given between the unseen action measure and the object measure.Moreover, we set nonuniform weights for both the actions and objects.The objects are weighted according to their transductive maximum score, with Z o a normalization constant over all objects in O s .
The unseen action weights are given as with Z a a normalization constant over all actions.The intuition behind the object weights is to focus the attention of the transductive universal transport on objects with a higher visual likelihood.The weights for the unseen actions are given as the inverse over the maximum word embedding similarity with respect to the objects, under the notion that actions without obvious relations to objects should have a more prominent spot in the transport coupling.With the optimal transport computed between unseen actions and objects, action prototypes are again interpolated following Equations 9 and 10, with the updated prototype for action label l now denoted as ω ‡ (l).
For the final zero-shot action inference from objects, we follow the same setup as current object-based approaches, where the score for each action label l in video v is determined based on the top relevant objects for that action (Jain et al, 2015;Mettes et al, 2021): with O l the set of most semantically similar objects for action label l.Finally, the transductive action-based and object-based scores from respectively Equations 11 and 15 can also be fused as s fusion (l|v) = ϵ•s action (l|v)+ (1 − ϵ) • s object (l|v).Summarized, our approach formulated for universal object models differs in three ways from its formulation to universal action models: the mapping is performed towards object embeddings rather than video embeddings, the object measure only includes objects that actually occur in any of the test videos, and the object measure is weighted based on video likelihood, where as the video measure is unweighted.

Datasets
Source datasets.We employ two source datasets for universal representation learning, namely Kinetics-700 for actions and ImageNet for objects.For Kinetics-700, we follow Brattoli et al (2020) and use a subset with 664 action categories to avoid any overlap with action categories in datasets used for zero-shot action recognition.For ImageNet, we follow Mettes et al (2021) and use the reorganized variant containing 12,988 object categories (Mettes et al, 2020).
Target datasets.The classification evaluation is performed on the two datasets used most often in zero-shot action recognition: UCF-101 and HMDB51.The UCF-101 dataset consists of 13,320 videos covering 101 action categories.Next to 101-way zero-shot evaluation, we also evaluate on settings with 20 and 50 test actions.For these settings, we rerun our approach on 10 runs with randomly selected actions and we report the mean and standard deviation over the runs.We note that in the 20-and 50-way zero-shot recognition, we do not use the other actions for training, they are simply not used in our approach.The HMDB51 dataset consists of 6,766 videos covering 51 action categories.Next to 51-way evaluation, we also investigate 10-and 25-way zero-shot recognition.
We also investigate the potential of our approach for zero-shot spatio-temporal action localization on UCF Sports and J-HMDB.UCF Sports consists of 150 videos with 10 actions and J-HMDB consists of 928 videos with 21 actions.For the evaluation, we follow Jain et al (2015) and report results with the AUC metric across five overlap thresholds.

Implementation details
We consider two universal action networks ϕ.First, we employ the R(2+1)D network (Tran et al, 2018) as given by Brattoli et al (2020), pre-trained on 664 Kinetics categories.For each video, we obtain its corresponding video embedding by randomly selecting a 16 frame shot and passing the shot through ϕ.For a fair comparison to Brattoli et al (2020), we also use the same word embedding ω, namely a word2vec model (Mikolov et al, 2013a), resulting in a 300-dimensional representation per word.For any action or object with more than one word, the word representations are averaged.Second, we employ the network of Pu et al (2022) with 25 splits per video, which is based on the same word embedding and Kinetics splits as the first action model.For both approaches, we rely on author-provided code to obtain action and video embeddings.For the universal object model, the object scores in a video are obtained following the protocol of Mettes et al (2021), where two frames per second are sampled, each fed to the pre-trained ImageNet model, and with the object probabilities averaged over the sampled frames.
For the clustering of the video embeddings, we use k-means clustering along with ℓ 2 -normalization, akin to Banerjee et al (2005).For the optimal transport, we employ the Lagrangian approach of Bonneel et al (2011) as implemented in (Flamary et al, 2021).Specifically, we set the cosine distance matrix as loss matrix, run for a maximum of 100,000 iterations if there has been no convergence and with the dual potential centered in the optimization.Unless specified otherwise, accuracy denotes the top 1 accuracy.Lastly for spatio-temporal localization, we start from the tubes made available by Mettes et al (2021).To each tube, we add the corresponding video-level action scores from our approach to improve the ranking of the action tubes over the entire dataset.The universal transport takes roughly 36 seconds CPU time for 13,320 videos, 101 actions, and 1000 clusters on UCF-101 on an Intel Xeon CPU.Once the action prototypes are re-positioned, no additional computational effort is required for zero-shot inference.All code will be made publicly available.Fig. 3: Evaluating universal prototype transport from seen actions on UCF-101.Left: The effect of the number of clusters and the interpolation ratio on the recognition performance.An interpolation rate of 1 denotes the baseline with the original unseen action prototypes.We find that re-positioning the prototypes directly boosts zero-shot performance given sufficient clusters, with a further boost by interpolating between the original and target prototypes, see the highest overall score for 1,000 clusters and interpolation ratio 0.5.Right: Intuition behind our improved results.Using the original unseen action prototypes results in large biases during zero-shot inference.With our approach, this imbalance is reduced, as indicated by the more even action distributions in the plot and the corresponding higher entropy scores in brackets in the legend.

Experimental results
We focus on five experiments: (i) evaluations on universal action models; (ii) evaluations on universal object models; (iii) integrating and fusing our approach with recent methods; (iv) state-of-the-art comparison for zero-shot action recognition and zero-shot spatiotemporal action localization; (v) qualitative analyses.

Universal transport from seen actions
Setup.For the first experiment, we evaluate on UCF-101 using all 101 actions for classification.We investigate the two variables that come with our approach in the context of universal action models, namely the granularity of the cluster aggregation over all test videos and the interpolation ratio between the original and target prototypes of the unseen actions.We use the universal action model of Brattoli et al (2020) throughout this experiment.
Results.The results for five cluster sizes and three interpolation ratios are shown in Figure 3a.An interpolation rate of 1 denotes the baseline using only the original action prototypes and 0 denotes the setting using the target prototypes.With the original semantic prototypes, we obtain an accuracy of 39.2%.Using the target prototypes directly boosts the classification accuracy when using sufficiently many clusters.Using only few clusters results in a coarse approximation of the distribution of test videos, which leads to lower performance.The best performance is obtained by positioning the unseen actions halfway along the geodesic between the original and target embeddings.With 1,000 clusters the accuracy becomes 42.4%, compared to 40.1% when using the target prototypes directly.We will use 1,000 clusters and an interpolation ratio of 0.5 for all other experiments involving universal action models.
Analysis.An explanation for our obtained improvements is shown in Figure 3b.We show the distributions of selected actions across all three interpolation ratios when using 1,000 clusters.With the original unseen action embeddings, this distribution is highly uneven, with 23% of the actions never being selected, naturally leading to zero accuracy for these actions.With our approach, the distributions become more uniform, highlighting the bias reduction.This is also reflected in the entropy of the action selection distributions in the top right of 3b, which increases when employing universal prototype transport, confirming that the distribution becomes more uniform.After performing our approach, the confusion matrix is more uniform and highlights a better performance for actions ignored by the baseline.
In Figure 4, we show the per-class performance gains as a function of selection frequency before our transport.The figure shows that our approach improves especially those classes that were not frequently occurring in a stand-alone universal model.This result highlights that our improvements are due to a better alignment between unseen action prototypes and projected test videos in semantic space.Examples of most improved actions include sky diving (accuracy from 8.2% to 80.9%), hammering (from ¡0.1% to 65.0%), and cricket bowling (from 0.0% to 55.4%).Figure 5 provides dives deeper into the observation by highlighting the improved alignment based on the entire confusion matrix.
As an extra test, we investigate the effect of class imbalance in the test set for universal prototype transport.1: Evaluating universal prototype transport from objects on UCF-101.Our proposed approach also enhances universal object-based approaches for zero-shot action recognition, especially when incorporating object filtering.
Following the long-tailed literature (Cui et al, 2019), we sample UCF-101 with exponential decays of factors 0.1 and 0.01.The average per-class accuracy on standard UCF-101 is 42.0% and remains stable (42.4% at imbalance ratio 0.1 and 41.6% at imbalance ratio 0.01), highlighting that our approach is stable to test-time class imbalance.

Universal transport from objects
Setup.Second, we investigate our approach on universal object models.We again use UCF-101 with all 101 actions for evaluation, with the interpolation ratio fixed to 0.5.We evaluate three threshold levels that come with the definition of objects as discrete measure, along with the universal object model itself and a vanilla uniformly-weighted optimal transport using all objects as baselines.
Results.In Table 1, we show the zero-shot action results for our approach when maintaining the top 2,500, 1,000, and 500 objects according to their transductive maximum likelihoods over all test videos.We first find that using a baseline optimal transport approach akin to the setup for seen actions provides only a marginal boost from 29.9% to 30.1%.In contrast, using the proposed weights for the unseen actions and objects, combined with a filtering of objects never present in a test video, provides a boost to 31.6% with the top 1,000 objects.We find that as long as the threshold is not set too strictly (e.g., keeping 1,000 objects or more) provides stable zero-shot results.
In Table 2, we show that the proposed weighting matters.For this Table we keep the top 1,000 objects and investigate all four combinations of uniform and proposed weighting.With uniform weights for both unseen actions and objects, the results are similar to the baseline object-based setup, while the results improve Table 2: Proposed versus uniform weighting between unseen actions and objects.Focusing on unseen actions with low semantic relation to any object (inverse) and on objects objects that are also observed in all test videos (transductive) improve the prototype transport from objects for zero-shot action recognition.
when incorporating either or both of the weight vectors to the proposed setup.We conclude that in the universal object-based model for zero-shot action recognition, the proposed transport is also beneficial.

Transporting multiple universal models
Setup.Third, we investigate how universal prototype transport operates across and in combination with multiple universal models.We focus on two results: providing an overview of our approach on multiple stateof-the-art universal models and combining action and object models with universal transport.
Results.In Table 3, we provide an overview of our approach on top of recent universal action and object models.For the universal action models of Brattoli et al (2020) and Pu et al (2022), we use the pre-trained models provided by the authors to compute video embeddings and add our universal prototype transport on top.
For the universal object model of Mettes et al (2021), we take the author-provided pre-trained object network and directly use the conventional object-to-action formulation of Equation 15 to obtain zero-shot predictions.We find that across multiple models, the results of both the top 1 and top 5 accuracy are improved.For action models, the top 5 improvements are even higher than the top 1 improvements, highlighting the better overall alignment between unseen actions and test videos.We conclude form this We show here that our transductive approach is general in nature and can be plugged on top of different universal models, improving both top 1 and top 5 accuracy.complementary nature.In Figure 6, we show the effect of combining both perspectives on UCF-101 and HMDB51.We fuse the action model of Brattoli et al (2020) with the object model of Mettes et al (2021) and leave the final fusion with the model of Pu et al (2022) for the state-of-the-art comparison.
On UCF-101, we find that fusing both approaches has a clear, positive effect.Our results improve from 42.4% (universal actions) and 31.6%(universal objects) to 47.9% when balancing both equally.When setting the fusion proportion to 0.3, the results can even be further improved to 48.9.Due to the zero-shot nature of our approach, we stick to an a priori equal balance between both setups.Averaged over all fusion ratios, our approach provides a boost of 3.0 percent point compared to the baseline fusion.On HMDB51 with universal action models, adding our approach improves the results from 24.9% to 28.1%.The results are further improved to 29.4% when fusing with universal object models.We conclude that both perspectives are complementary for zero-shot action recognition and their fusion benefits from our proposed transport.

Comparison to state-of-the-art
Zero-shot recognition.In Table 4, we compare our results on UCF-101 and HMDB51 to the state-of-theart in zero-shot action recognition.Similar to other universal approaches, in the scenarios with random subselection of the test actions (20 and 50 for UCF-101, 10 and 25 for HMDB51) we do not use the remaining actions and their videos for network training.On both datasets, the current state-of-the-art is given by Lin et al (2022) and Pu et al (2022).For the action number of our approach, we add our universal transport on top of the two action models used in Table 3.We note that when reproducing the results of Pu et al (2022) Fig. 6: Fusing action and object information for zero-shot recognition on UCF-101 and HMDB51.Combining universal action and object information benefits zero-shot recognition, with universal prototype transport preferred across all fusion proportion between both sources.Table 5: State-of-the-art comparison on UCF Sports and J-HMDB for localization with five overlap thresholds.Across datasets and thresholds, we obtain the highest scores, highlighting the effectiveness of our approach in the context of zero-shot spatio-temporal localization.Fig. 7: Qualitative examples from UCF-101.Columns 1-3: Success and failure cases of our approach on the universal action model of Brattoli et al (2020).Our approach can help to better align videos with unseen actions, as shown in the skydiving and nunchuck videos, but can fail for actions such as front crawl due to confusion with a similar action like breast stroke.Columns 4-5: Qualitative examples from the fusion of universal action and object models in our approach.In both cases, we are better able to classify fine-grained unseen actions by transferring dynamic knowledge about other actions and static knowledge about the objects used to perform these actions.
Table 4 details comparisons with both complete testsets and smaller subsets.Averaged over the multiple runs, our approach is effective for both small and large dataset sizes.Compared to the baseline models used as starting point in our method, we find that the larger the testset, the higher the relative improvement, indicating that our approach benefits from richer semantics.
Zero-shot localization.We also showcase the potential of our approach for zero-shot spatio-temporal action localization.Since our approach operates over entire videos, we start from the spatio-temporal tubes made publicly available by Mettes et al (2021).For each tube, we simply add the score for each action from the entire video as given by our approach.In Table 5, we report the AUC scores for UCF Sports and J-HMDB.Across datasets and overlap thresholds, we find that the global scores from our approach boosts spatio-temporal localization.This is because the scores of our approach help to distinguish and rank tubes from different videos as they encode contextual information.
In conclusion, we find that for both zero-shot classification and spatio-temporal localization on all datasets, our approach provides consistent improvements, highlighting that universal prototype transport is effective across different collections of unseen actions.

Qualitative analysis
In Figure 7, we show success and failure cases for our approach when applied to the universal action model of Brattoli et al (2020) and its fusion with the universal object model of Mettes et al (2021).Theses results reiterate the potential of combining action and object perspectives in zero-shot action recognition and the role of universal prototype transport in combining both views.

Conclusions
In this work, we investigate a persistent limitation in current universal learning models for zero-shot action recognition, namely selection biases in the assignment of unseen actions to test videos.We introduce universal prototype transport to alleviate this limitation.Our approach consists of three stages: (i) finding an optimal transport mapping from unseen action prototypes to the projected test videos (in universal action models) or to object prototypes (in universal object models); (ii) obtaining a target prototype for each unseen action using the couplings from the hyperspherical optimal transport; and (iii) re-positioning the unseen actions along the geodesic spanned by the original and target prototypes.Empirical evaluation on four datasets shows the effectiveness of our approach for debiasing action assignments and for improving zero-shot recognition and localization as a result.Our approach is general and can be used to improve any universal model.

Fig. 4 :
Fig.4: Per-class improvements as a function of the number of times an action is selected prior to universal prototype transport on UCF-101.Our approach improves especially those classes that are rarely selected in a stand-alone universal action model.

Fig. 5 :
Fig. 5: Confusion matrices on UCF-101 before (left) and after (right) performing universal prototype transport on the model by Brattoli et al (2020), sorted by selection frequency in the base model.For the baseline, the entire right side of the matrix is dark blue, since those actions are never selected by any test video.After performing our approach, the confusion matrix is more uniform and highlights a better performance for actions ignored by the baseline. Table

Table 3 :
Universal prototype transport on top of several inductive action and object models on UCF-101.

Table 4 :
(Zhuo et al, 2022)omparison on UCF-101 and HMDB51 for different numbers of test actions.On all settings except UCF-101 with 50 classes, our approach obtains the highest zero-shot action classification scores and we expect further improvements when using Swin Transformers as employed by(Zhuo et al, 2022).