MambaNet: A Hybrid Neural Network for Predicting the NBA Playoffs

In this paper, we present Mambanet: a hybrid neural network for predicting the outcomes of Basketball games. Contrary to other studies, which focus primarily on season games, this study investigates playoff games. MambaNet is a hybrid neural network architecture that processes a time series of teams' and players' game statistics and generates the probability of a team winning or losing an NBA playoff match. In our approach, we utilize Feature Imitating Networks to provide latent signal-processing feature representations of game statistics to further process with convolutional, recurrent, and dense neural layers. Three experiments using six different datasets are conducted to evaluate the performance and generalizability of our architecture against a wide range of previous studies. Our final method successfully predicted the AUC from 0.72 to 0.82, beating the best-performing baseline models by a considerable margin.


INTRODUCTION
Sporting events are a popular source of entertainment, with immense interest from the general public.Sports analysts, coaching staff, franchises, and fans alike all seek to forecast winners and losers in upcoming sports match-ups based on previous records.The interest in predicting sporting outcomes is particularly pronounced for professional team sport leagues including the Major League Baseball (MLB), the National Football League (NFL), the National Hockey League (NHL), and the National Basketball Association (NBA); postseason plays in these leagues, namely the playoffs, are are of greater interest than games in the regular season because teams compete directly for prestigious championships titles.
The development of statistical models to robustly predict the outcome of playoff games from year-to-year is a challenging machine learning task because of the plethura of individual, team and extenral factors that all-together confound the propensity of a given team to win a given game in a give year.
In this work, we develop MambaNet: a large hybrid neural network for predicting the outcome of a basketball match during the playoffs.There are five main differences between our work and previous studies: (1) we use a combination of both player and team statistics; (2) we account for the evolution in player and team statistics over time using a signal processing approach; (3) we utilize Feature Imitating Networks (FINs) [1] to embed feature representations into the network; (4) we predict the outcome of playoff results, as opposed to season games; and (5) we test the generalizability of our model across two distinct national basketball leagues.To assess the value of our proposed approach, we performed three experiments that compare MambaNet's to previouslyproposed machine learning algorithms using NBA and Iranian Super League data.

RELATED WORK
The NBA is the most popular contemporary basketball league [2,3].Several previous studies have examined the impact of different game statistics on a team's propensity to win or lose a game [4,5].More specifically, previous studies have identified teams' defensive rebounds, field goal percentage, and assists as crucial contributing factors to succeeding in a basketball game [6]; for machine learning workflows, these game attributes may be used as valuable input features to predict the outcome of a given basketball game [7,8].
Probabilistic models to predict the outcome of basketball games have been proposed by several previous studies.Jain and Kaur [9] developed a Support Vector Machine (SVM) and a Hybrid Fuzzy-SVM model (HFSVM) and reported 86.21% and 88.26% accuracy in predicting the outcome of basketball games.More recently, Houde [10] experimented with SVM, Gaussian Naive Bayes (GNB), Random Forest (RF) Classifier, K Neighbors Classifier (KNN), Logistic Regression (LR), and XGBoost Classifier (XGB) over fifteen game statistics across the last ten games of both home and away teams.They also experimented over a more extended period of NBA season data, starting from 2018 to 2021, and reported 65.1% accuracy in winners/losers classifications.In contrast to Kaur and Houde, that addressed game outcome prediction Fig. 1: An overview of MambaNet's architecture.First, the home (column 1, yellow boxes) and away (column 1, purple boxes) teams' stats and the two teams' players' stats are fed to the network.Next, four FINs are utilized to represent the input stats' signal features, which contain trainable (column 1, dark circles) and non-trainable (column 2, light circles) layers.These representations are further processed with convolutional and Dense layers.Raw time-domain signal features are also extracted from input stats using LSTM networks.Finally, the aforementioned features are incorporated to make the final prediction.as a binary classification task, Chen et al [11] identified the winner/loser by predicting their exact final game scores.They used a data mining approach, experimenting with 13 NBA game statistics from the 2018-2019 season.After feature selection, this number shrank to 6 critical basketball statistics for predicting the outcome.In terms of classifiers, the authors experimented with KNN, XGB, Stochastic Gradient Boosting (SGB), Multivariate Adaptive Regression Splines (MARS), and Extreme Learning Machine (ELM) to train and classify the winner of NBA matchups.The authors also studied the effect of different game-lag values (from 1 to 6) on the success of their utilized classifiers and indicated that 4 was found to perform best on their feature set.
Fewer studies have used Neural Networks to predict the outcome of basketball games; this is mostly due to challenges of over-fitting in the presence of (relatively) small basketball training datasets.Thabtah et al [12] trained Artificial Neural Networks (ANN) on a wide span of data where they extracted 20 team stats per NBA matchup played from 1980 to 2017.Their model obtained 83% accuracy in predicting NBA game outcomes; they also demonstrated the significance of three-point percentage, free throws made, and total rebounds as features that enhanced their model's accuracy rate.

METHODS
Baseline approach: A majority of the existing studies use a similar methodological approach: For each team (home and away), a set of s game statistics (the features) are extracted over n previous games (the game-lag value [11]) forming an n × s matrix.Then, the mean of each stat is calculated across the n games, resulting in a 1 × s feature vector for each team.The two feature vectors and concatenated yielding a 1 × 2s vector for each unique matchup between a given pair of teams.Finally, this results in a trainSize × 2s matrix which is used to train classification model (each experiment will report the train/set set size in more detail).Alternatively, the label of each sample indicates whether the home team won (y = 1) or lost the game (y = 0).FIN Training: Our method follows the same steps as the baseline approaches, but with one critical difference: instead of calculating the mean of features across the n last games using the mean equation, we feed the entire n × s matrix to a pretrained mean FIN and stack hidden layers on top of it (hereafter, this FIN-based deep feedforward architecture is referred to as FINDFF) to perform binary classification; In addition to the mean feature, we also imitate standard deviation, variance, and skewness.All FINs are trained using the same neural architecture: A sequence of dense layers with 64, 32, 16, 8, and 4 units are stacked, respectively, before connecting to a single-unit sigmoid layer.The activation function of is ReLU, for the first two hidden layers and the rest are Linear.Each model is trained in a regression setting by using 100,000 randomly generated signals as the training set and handcrafted feature values for each signal as the training labels.Then, we freeze the first three layers, finetune the fourth layer, and remove the remaining two layers before integrating them within the larger network structure of the MambaNet network.

Mambanet:
In Figure 1, we provide an illustration of Mambanet -our proposed approach.The complete set of player and team statistics used in this study can also be found in Table 1.The input to the network is an 10×35 stats matrix which are passed to both the pretrained FINs as well as LSTM layers to extract a team's statistics' sequential features.For each team, we also extract the a stats matrix (n = 10, s = 34) for each of its roster's top ten players and pass them to the same FINs and LSTM layers.Next, we flatten teams' signal feature representations and feed them to dense layers, whereas for players, we stack them and feed them to 1D convolutional layers.Finally, all latent representations of a team and its ten players are concatenated in the network before connecting them to the last sigmoid layer.

EXPERIMENTS & RESULTS
We performed three experiments to assess the performance of our proposed method.To demonstrate the advantage of leveraging FINs in deep neural networks, we first compare the performance of FINDFF against a diverse set of other basketball game outcome prediction models trained using NBA data.For the second experiment, these models are tested for generalization across unseen basketball playoff games from the Iranian Super League data.Finally, we assess the performance of Mamabanet for accurate playoff outcome prediction.In all three experiments, the Area Under the ROC Curve (AUC) was used as our primary evaluation metric.

CONCLUSION
In this work, we tackled playoff basketball game outcome prediction from a signal processing standpoint.We introduced MambaNet, which incorporated historical player and team statistics and represented them through signal feature imitation using FINs.To compare our method with the baseline, we used NBA and Iranian Super League data which enabled us to demonstrate the performance and generalizability of our method.Future studies will potentially use fusion techniques or other suitable data modeling techniques, such as graphs, to develop more advanced neural networks that integrate team and player representations more efficiently to predict playoff outcomes more accurately.

Table 1 :
A description of game statistics used in this work.Except for the last three features, the rest (1 to 33) are shared statistics in representing both teams and players.(Abr: Abbreviation, #: feature number)

Table 3 :
A performance comparison between FINDFF and other previouslydeveloped machine learning models trained on five years of NBA (from 17-18 to 21-22) on the 2020-2021 Iranian Super League Playoffs (Ref: Reference, FC: Feature Count, Alg: Algorithm, TW: This Work)

Table 4 :
Comparing the performance of different MambaNet versions in In Table4, we present the results of our incremental experiment.The first row reports the simplest version of MambaNet using 35 team features that are passed to a FINDFF network imitating the mean (m) as a feature.Compared with the baseline, we use a more extensive set of basketball game statistics to form the feature vector of a team since this helps better satisfy the data-intensive requirement of neural networks.At this stage, the AUC varies between 0.70 to 0.72.Next, we trained three more FINs to imitate Standard Deviation std, Variance v, and Skewness s using the same neural network architecture as the mean FIN.The second row of the table shows how adding new signal feature representations improves the AUC up to 0.10 and 0.03 in the 2018-2019 and 2020-2021 NBA datasets.Furthermore, we integrated players' statistics alongside team statistics, leading to an 0.02 increase in AUC across four NBA datasets in the third row.Lastly, as shown in the fourth row, we used RNN layers to create a time-series representation of the game and individual statistics, resulting in 0.03 and 0.02 improvements in 2019-2020 and 2021-2022, respectively.