1 Introduction

Isaac Asimov’s science fiction ‘Foundation’ trilogy is based on the presumption that it may be possible to develop a framework which would allow to model and predict the political and economic development of human societies as such [7, 31]. At the core of the framework in question, ‘psychohistory’, lies the hypothesis that it could be possible to describe collective decision-making by rigorous laws, regardless of what a specific individual decides to do. Individual preferences would average out when populations are large. On a decisively less galactic level, a legitimate question regards the statistics of cultural phenomena produced by large numbers of decisions. Prime candidates are in this regard rankings measuring the economic success of cultural goods, like music albums, books and movies. Here we will show that certain traits of the New York Times best-seller list and the streaming charts compiled on a daily basis by Spotify, a streaming portal, are congruent with the predictions of an information-theoretical theory of human decision-making.

It has been recognized that charts, in particular music charts, like the US-based Billboard charts, provide access to the study of an extended range of socio-cultural developments. Examples are the evolution of harmonic and timbral properties of popular music [21], the influence of gender and race on chart success [16], and how technological progress influences cultural diversity, respectively, concentration processes [29]. Other studies investigated the correlation between acoustic features [15] and repetitive lyrics [28] to market success, the interplay between popularity and significance in popular music [23], and whether there is a trend towards a global musical culture [1]. An alternative venue for the study of long-term societal developments are recognized scientific and cultural awards, like the Nobel Prize [10].

On an extended level, Schneider and Gros [34] suggested that music charts are suitable proxies for testing quantitatively the concept of social acceleration [33], viz. the notion that both personal and societal processes proceed nowadays faster than they used to [38, 39]. The rationale for this suggestion is based on the observation that long-term databases of comparable entries are indispensable for studies aiming to measure the evolution of culturally and sociologically relevant time scales, a condition satisfied by charts, music and book charts alike, which provide timelines of five and more decades of weekly compiled data. For the US, UK, German and Dutch music charts it was found [34], that the evolution of the album lifetimes and of the weekly rank decay provides evidence for a substantial acceleration, in particular since the rise of the internet. These findings are congruent with the changes observed for the chart diversity and for the dynamics of number-one hits.

Cultural items, like books and music albums, compete for resources, such as monetary budgets, attention and dedication. Competition for time and attention is also a driver for the dynamics of news topics on social networks, a process that has been argued to take place on accelerating time scales [18]. The success of social media generated content is closely connected to sharing, which is an aspect not present for classical cultural items. It has been observed in this context, that the statistics of sharing and scientific citations seem to follow identical universal functionalities [25].

An accessible example of a social platform is Reddit [22], a user generated exchange platform for information, opinions and comments. Studies of Reddit data examined [2], among other things, the relation between popularity and intrinsic quality [36], the stability of online communities with substantial turnover rates [30], and the evolution of user interests [37].

Taking the number of comments as a popularity measure, one can generate charts which parallel download and sales charts. One key difference is however that the exact timing of the triggering event is now of relevance. Download charts contain albums that have been released not only the same day, but also considerably in the past. In contrast, new posts on Reddit remain popular only for a maximum of 1 or 2 days, which implies that the day–night activity cycle of the user could impact the lifetime distribution of Reddit posts. This is indeed what one finds. We argue that this observation implies that the feedback between popularity rankings and user activity interferes with additional timescale, like the 24 h day–night activity cycle, when the timing of the user activity is strongly correlated with that of the triggering event.

Book and music sales and streaming charts are primarily consumption charts, for which an aggregate information-theoretical theory is proposed. Exactly when a consumption takes place, like the acquisition or the downloading of books and music, is not relevant, in contrast to the posting of Reddit comments, and other social media activities, which are time-relevant because they involve the creation of new items. As a consequence, we find when analyzing Reddit and Twitter comment and sharing charts, that additional driving forces, in particular the 24 h day–night cycle, tend to mask the development of fully autonomous self-organization for charts of social media activities.

2 Information theory of human decision-making

On a statistical level, human decision-making may be modeled using an information theoretical ansatz [11]. Starting from the same basic ansatz as Schneider and Gros [34], we present here a modified derivation. From an epistemology view, we stress that our ansatz is intended as a possible explanatory framework. Similarly, the data analysis presented further below is not claimed to prove the correctness of the information theoretical ansatz, indicating, however, that it is a valid contender. From a generalized perspective our ansatz is neutral [17], in the sense that is based not on semantic, but on statistical considerations.

2.1 Decision-making maximizing information

The brain processes and stores information. Given the finite size, information needs to be selected and compressed [20]. A suitable measure for the amount of information encoded by the probability distribution p(s) of a quantity s is given by the Shannon Entropy H[p],

$$\begin{aligned} H[p] = -\langle \log p\rangle , \quad \langle A\rangle = \int \mathrm{d}s\, A(s)\,p(s),\quad \end{aligned}$$
(1)

where \(\langle A\rangle \) denotes the expectation value of the function \(A=A(s)\), see, e.g. Gros [8]. We assume that the overall goal of human decisions is to maximize information, in particular the information generated by decision-making processes. As a proxy for the impact of actions, e.g. to buy a cultural item, we examine the statistical properties of the corresponding chart.

Statistically, information maximization is equivalent to maximizing the entropy H[p]. This goal can be achieved in practice only when respecting a given set of constraints. Real-world actions typically need to factor in the amount of effort and time involved, as well as the uncertainty of the outcome, viz. the variance. On a statistical level these two constraints are given by the mean and the variance, which are equivalent to the first and the second moment of the probability distribution in question, \(\langle s\rangle \) and \(\langle s^2\rangle \). The constrained maximum entropy distribution is consequently obtained by maximizing the objective function:

$$\begin{aligned} \Phi [p] = H[p] - a \langle s\rangle - b \langle s^2\rangle , \end{aligned}$$
(2)

where a and b are suitable Lagrange multipliers. Using standard variational calculus [8], the distribution function maximizing \(\Phi [p]\) is found to be a Gaussian with mean \(\mu =-a/(2b)\) and variance \(\sigma ^2=1/(2b)\),

$$\begin{aligned} p(s) \sim \mathrm {e}^{-a s-b s^2} \sim \mathrm {e}^{-(s-\mu )^2/(2\sigma ^2)}, \end{aligned}$$
(3)

where normalization factors have been suppressed.

Table 1 Parameters of lifetime distributions. Listed are the linear and quadratic parameters a and b from Eq. (5). Note that \(a+1\) corresponds to the exponent of the power law contribution, see (6). The corresponding quadratic curves in log-log space are shown in Figs. 2 and 3. The table is sorted with respect to the linear parameter a. The uncertainties refer to the margin of error given a 95% confidence level

2.2 Human brains compress information logarithmically

When making decisions, which is the distribution function p(s) that is to be maximized statistically using the entropy H[p]? For the case of book and music charts, a prime candidate is the lifetime distribution p(L), which measures the probability that a given item, a book or a music album, remains listed for a period L, the lifetime. While listed, books and albums receive increased attention, which implies that the listing period constitutes a visible effect of the individual decision to buy a specific cultural item.

From the cognitive perspective, the brain is confronted with two contrasting demands: to store incoming information as faithfully as possible, covering at the same time the extended orders of magnitude characterizing physical stimuli, like sound and light intensity, as well as time scales. An efficient solution for this conundrum is to store information on a compressed scale, for instance logarithmically. This is indeed the case, as expressed by the Weber–Fechner law, which states that the brain discounts sensory stimuli [13], numbers [6, 27] and time logarithmically [14]. It is not a coincidence that we use logarithmic scales, lumen and decibel, to measure light and sound intensities.

2.3 Maximizing compressed information

When performing an operation, like information maximization, the brain uses internal states that we denote here s. These states are in general related logarithmically to outside quantities, as discussed in the previous section. For the case of the lifetime distribution p(L), we have

$$\begin{aligned} s=\log (L), \qquad \quad p(s)\mathrm{d}s = p(L)\mathrm{d}L, \end{aligned}$$
(4)

which relates the observable distribution function p(L) with its internal representation, the probability density p(s). Using \(\mathrm{d}s/\mathrm{d}L=1/L\) and the maximum entropy probability density p(s), as given by (3), one finds

$$\begin{aligned} p(L) \sim \frac{1}{L}\mathrm {e}^{-a \log (L)-b \log ^2(L)} \sim \mathrm {e}^{-(\log (L)-\tilde{\mu })^2/(2\tilde{\sigma }^2)},\nonumber \\ \end{aligned}$$
(5)

with a log-mean \(\tilde{\mu }=-(a\!+\!1)/(2b)\), and a log-variance of \(\tilde{\sigma }^2=1/(2b)\). p(L) corresponds to a log-normal distribution whenever \(b>0\), and to a power law,

$$\begin{aligned} p(L) \sim \frac{1}{(L)^{a+1}}, \qquad \quad b\rightarrow 0, \end{aligned}$$
(6)

if b vanishes. Power laws, which are frequently observed [19], are hence a natural outcome of human activities, statistically, whenever the variance is either small or not taken into account.

There are two causalities for why the variance may be of secondary importance. First, when uncertainties are small, viz when fluctuations around the mean are negligible. In this case the b-term in (5) is numerically small, becoming important only for exceedingly large or small lifetimes L. A second, more general argument is that one has to invest comparatively more resources to sample the variance than just the mean, in particular when correlations are present [4]. If time is a scarce resource, only the mean can be obtained reliably from a time dependent functionality, such as the lifetime distribution of cultural items. As a consequence it follows, that power laws are present when individual decision times are shorter than typical chart listing duration, with log-normal distributions emerging for extended decision times.

The probability density p(L) is normalizable only for \(b\ge 0\) if arbitrary large lifetimes L are allowed. Real-life time intervals are however finite, which makes a negative b also viable. The values for \(a+1\) and b obtained throughout this study for different music and social network activity charts are discussed in the next section.

Fig. 1
figure 1

Long-term evolution of charts and best-seller lists. Left: The chart diversity \(d=N_\mathrm{a}/N_\mathrm{s}\), with \(N_\mathrm{a}\) denoting the number of unique titles on a chart with \(N_\mathrm{s}\) slots. The chart diversity of the New York Times best-seller list (orange) follows a trend similar to the top 10 of the Billboard album chart (purple). The top 100 album chart (blue) is also shown for comparison. The dotted lines indicate the inclusion of streaming data into the Billboard album chart (see Sect. 4.1 for details). Right: Out of all number-one best sellers/hits, the fraction \(P_\mathrm{one}\) of books/albums that top the best-seller list/charts already on entry, viz. the first time they are listed. Initial success is all important for today’s cultural items

3 Results

The information-theoretical approach to the statistics of human decision-making implies a characteristic functional dependency for the distribution of the lifetime of cultural items. We tested this dependency for book and album sales charts, for music download charts, as well as for social media comments. Data retrieval and processing is described in Sect. 4. To visualize the possible p(L) distribution functions, figures with log-log representations of lifetime distributions include quadratic fits that correspond to Eq. (3). However, these fits are not claimed to correspond to statistically validated models, as discussed, e.g., by [5]. In Tables 1 and 2, the respective parameters are listed.

3.1 Chart diversity

An intuitive measure for the turnover rate of charts is the chart diversity \(d = N_\mathrm{a}/N_\mathrm{s}\), where \(N_\mathrm{a}\) is the number of unique titles, here per year, normalized by the number of available slots \(N_\mathrm{s}\) (for the whole year). Figure 1 shows the evolution of the chart diversity of the New York Times best-seller list, in comparison with the top 10 and top 100 Billboard album charts. Billboard introduced streaming data into the chart in 2014 but still published a sales chart. This is why the line splits in 2014, with the chart including streaming shown as a dotted line.

The chart diversities of the New York Times best-seller list and of the Billboard album charts have taken a similar evolution over the course of the last three to four decades. This fact is quite remarkable, given that individual music albums and fiction books differ substantially concerning their respective consumption times. Generally one needs substantially longer time to read a novel than to listen to an album. Their overall key core characteristics are on the other hand similar. The average number of weeks a novel spent on the New York Times best-seller list \({\bar{w}}\approx 1/d\) (where d is the chart diversity), shrunk from \({\bar{w}} \approx 1/0.1 = 10\) weeks in the 1980s to about \({\bar{w}} \approx 1/0.3 \approx 3 \) in 2010. The equivalent numbers for the top-10 Billboard album charts are \({\bar{w}} \approx 1/0.1 = 10\) in the 1980s and \({\bar{w}} \approx 1/0.4 = 2.5\) in 2010.

Table 2 New York Times best-seller lifetime distribution parameters. Listed are the linear and quadratic parameters a and b, respectively, for the log-normal distribution Eq. (5), as shown in Fig. 4. Note that the prefactor of the linear term is \(a+1\), see (6). The uncertainties refer to the margin of error given a 95% confidence level

The New York Times best-seller list and Billboard album charts are also similar with respect to the dynamics of titles that eventually make it to the top, the number-one titles. For this there are two possible courses. Either a title enters the chart directly at the top, then its number-one position is guaranteed. Alternatively a given title starts lower and works its way up to the top, over the course of several weeks or months. The probability for a number-one title to take the first course of action, \(P_\mathrm{one}\), is shown in Fig. 1. In the 1970s and 1980s close to no number-one best seller entered the New York Times best-seller list at the top. This changed dramatically in the following three to four decades. From the late 2000s onward close to all, about \(95\%\) of all number-one novels started charting right at the top. This is similar to the evolution the Billboard album charts took. A small difference is that the New York Times best-seller list started the transition about five years prior to the Billboard album chart, however with the evolution taking about ten years longer. In general the Billboard top 10 and top 100 charts differ only marginally.

The changes in \(P_\mathrm{one}\) are also reflected in the time it takes on the average to become number one. Before 1980, when close to all number-one titles had to work their way up to the top, this process took on the average 5–8 weeks. Today number-one novels as well as albums become number one in under a week on the average, since almost all number-one titles start as such.

Fig. 2
figure 2

Lifetime distribution of music albums. The probability of album lifetimes, the number of weeks listed, for the Billboard (left) and Spotify (right) charts. A log-log representation has been used, for which log-normal distributions correspond to parabolas and power laws to straight lines, see Eq. (3). Two three year periods, 1980–1982 (orange) and 2017–2019 (yellow) are compared for the Billboard data. Daily (violet) and weekly (blue) Spotify album charts have been evaluated for 2017–2019. Note the outlier for the Spotify daily charts at a lifetime of exactly 1 week, which coincides to the minimal lifetime of the weekly chart. The Billboard charts are consistently compiled on a weekly basis

3.2 Musical chart lifetimes

In Fig. 2, the lifetime distributions of daily and weekly music album charts are presented, namely for the Billboard album charts, which are published on a weekly basis, and for the respective Spotify download charts.

Comparing the functional dependency of past (1980–1982) and present day (2017–2019) Billboard album lifetimes, one observes a transition from a log-normal distribution to a power law, as parameterized by (3). An equivalent transition is seen between daily and weekly Spotify charts. A caveat is here the outlier that can be observed for the Spotify daily charts at a lifetime of exactly 1 week. The origin of this outlier is unknown, it may be caused possibly by weekly algorithmic influences, such as maintenance periods. It is in any case remarkable, that the lifetime distributions presented in Fig. 2 change on a functional level, either with time or when the charting period is increased. We argue, that a separation of time scales causes both transitions.

Fig. 3
figure 3

Spotify single downloads and Twitter hashtag popularity. Left: Lifetime distribution of individual songs on Spotify daily and weekly single charts. Data for 2017–2019. Right: For top-50 Twitter hourly, daily and weekly hashtag charts the lifetime distribution. A kink corresponding to 24 h is present for the 1 h charts. Dashed/solid lines for the hourly charts correspond to quadratic fits to the data below/above the kink. Data from Mønsted [24]

Traditionally, music albums were bought in music shops, which involved a personal trip, and hence a comparatively long execution time. Long decision and execution times imply, as argued in Sect. 2.3, that the variance of the album lifetime distribution is relevant, which takes consequently the form of a log-normal distribution. The typical time needed to acquire a music album dropped, however, substantially below 1 week when online shopping started to become relevant in the 1990s. Given that 1 week corresponds to the chart frequency, this development implies that the variance of the album lifetime distribution lost its relevance. The desire to maximize information, the basis of the here discussed theory of statistical decision-making, leads in this case to power laws.

We believe that an analogous argument explains why daily and weekly Spotify album charts show respectively log-normal and power law distributions. On the average, it is presumably not a matter of only a few hours to listen and to appreciate music albums, which may contain a substantial number of titles, but of a few days. This time scale, several days, lies between the frequencies of daily and weekly streaming charts, which would explain why daily and weekly album charts are respectively log-normal and power law distributed.

Whereas the Billboard album charts are sales-based, the respective single charts are compiled predominantly on the basis of airtime statistics, for which the number of times songs are played by radio stations are counted. A direct comparison of album and single charts is in this case not possible [34]. The equivalent caveat does not hold for the Spotify charts, for which both album and single charts are based on streaming counts.

In Fig. 3, the lifetime distributions of daily and weekly Spotify single charts are shown. Both are close to power law distributions with only small quadratic contributions. The overall process, to listen to a music sample, to decide to stream, and to actually do it, can be assumed to take only a few minutes for individual songs, but substantially longer in the case of albums. The observation that already daily single download charts show indications for power law behavior, with respect to the distribution of song lifetimes, is hence consistent with our basic framework.

3.3 New York Times book best sellers

In Fig. 4, we present the probability that a given title stays for a certain number of weeks on the fiction best-seller list of the New York Times. In contrast to the top-100 music charts discussed further above, only the top ten titles of the week are included in the NYT best-seller list. To compensate for the comparatively limited database, we averaged the lifetime distribution over consecutive 10-year periods. Substantial scattering of the data can be observed nevertheless.

Included in Fig. 4 are quadratic fits, \(-(a+1)s-bs^2\), to the log-log representation of the lifetime distribution (compare Eq. (3)). It is notable that the distribution of book lifetimes evolves from being convex in log-log scale in the 50s and 60s to be being concave, starting from the 80s. The turning point, in the 70s, matches roughly a shallow minimum in the chart diversity, as shown in Fig. 1. It is presently unclear what drives this interesting phenomenon. The presence of a finite second-order component, \(b\ne 0\), indicates in any case that the individual time scales for buying and reading books has not dropped below 1 week, the charting frequency. A power law appears in the 70s, when the lifetime distribution transits from convex to concave. The power laws seen in the distribution of musical charts emerge in contrast in the concave region, see Fig. 2, when an initial \(b>0\) progressively vanishes [34]. This indicates two distinct causalities for the two power laws, for the late-state Billboard lifetime distributions and for the book lifetimes in the 70s.

Fig. 4
figure 4

Six decades of NYT best seller lifetimes. The lifetime distribution of books on the New York Times best-seller list. Shown is the average over 10-year periods. Note that only ten books are listed weekly, the data are hence comparatively sparse. In a log-log presentation, as shown, a long-term evolution from convex to concave is observed

3.4 Dynamics of Reddit comments

New data items, like posts or images, are at the core of social media activity. The response of other users, in the form of comments or likes, will determine the popularity of social media posts and with this the likelihood of additional likes and comments. In this context, we analyzed the statistics of Reddit post commenting, using publicly available data [32]. Note that Reddit is a discussion website organized around collections of user-created discussion boards for individual subjects called “subreddits”. Users can submit posts such as text posts, links, images and videos to a subreddit. Posts are then voted up or down and commented on. While upvoting can be done directly on the post overview screen while scrolling through, to comment, users have to actively click on the post.

The most popular posts from each subreddit are also shown on the front page on login. There are multiple options for sorting to determine which posts are the most popular. The old default was “Hot”, which was based on the number of upvotes (log-weighted) and the submission time (recent posts prioritized). In 2018, “Best” was introduced as the new default, which tries in addition to show new content, including weighting based on comment activity. Comment data is publicly available [32], including precise timestamps, vote counts are available however only as monthly averages. The technical aspects of our data analysis is described in Sect. 4.3.

For book and album charts the ranking criteria are the number of sales and downloads within the respective chart periods. In analogy, we used the number of comments for Reddit posts. In Fig. 5, the chart diversity and the respective lifetimes for 10- and 60-min Reddit top-100 charts are shown. Within the time window analyzed, 2013–2015, a downward trend in chart diversity is observed, with an equivalent increase in chart lifetimes. This is in contrast to the long-term trend for book and album charts that can be observed in Fig. 1.

The distributions for 10- and 60-min chart lifetimes presented in Fig. 5 is non-monotonic. On a coarse level, the probability to observe a certain lifetime drops in a power law-like fashion, which is, however, interseeded by a pronounced local maximum. The maxima correspond to characteristic time scales of about 20–26 h, for both the 10- and the 60-min charts, which suggests that the intrinsic 24 h day–night activity cycle may be involved. Users may sleep on a post or comment of the day, to revisit it in the next morning. A lesson learned from the Reddit database is then, that the presence of characteristic timescales may interfere strongly with the otherwise operative feedback loop between posts and comments.

3.5 Twitter hashtags

We use a publicly available corpus of Twitter hashtag statistics to study the residence time of hashtags in hourly, daily and weekly top-50 charts. See Sect. 4.4 for data source and handling. Hashtags and Reddit comments are both examples of active user contributions, albeit with a key difference. Adding a hashtag to a tweet can be presumed to be on the average less time-consuming than composing an entire Reddit comment. We may hence expect the trend towards self-organization to be more pronounced for Twitter hashtag charts, than for Reddit comments.

In Fig. 3, the distribution of Twitter hashtags for hourly, daily and weekly top-50 charts are presented. With respect to the Reddit data shown in Fig. 5 one observes that the 24 h activity cycle is now substantially less pronounced. Instead of a peak, the dominant feature of the lifetime distribution of the hourly Twitter hashtag charts is a kink. To be precise, only a weak local maximum occurs at about 18 h for the hourly charts, and a 30% drop at 24 h. For daily and weekly charts no anomaly is observed. Fitting the Twitter hashtag data with the maximum entropy distribution (3), we decided to split the 1 h lifetime data in two parts, below and above the above discussed kink. Within the log-log representation, one finds that the quadratic contribution tends to become smaller for longer charting periods, viz. when going from 1 h to daily and weekly charts. This observation complies with the argumentation laid out in Sect. 2.3, namely that lifetime distributions become more power law-like when the time scale of the individual activities is substantially smaller than the charting periods.

Fig. 5
figure 5

Reddit comment charts. Top: The evolution of the diversity of 10/60 min comment charts over the observation period (left) and the respective average lifetimes (right). Bottom: Respectively for the first six months of 2013 and the second half of 2015, the distributions of comment lifetimes, both for 10 min and 60 min charts. Adaptive binning has been used, as described in Sect. 4.5, with either a minimum of 70 or 1000 comments per bin. The local maxima in the lifetime distribution (vertical arrows) correspond to 1240/1560 min (20.7/26.0 h) for the 2013 10/60 min charts and to 1160/1320 min (19.4/22.0 h) for the 2015 10/60 min charts. See Sect. 4.3 for data processing

4 Data sources and processing

4.1 Book and album charts

The New York Times book best sellers [26], the Spotify music streaming charts [35] and the Billboard sales-based music charts [3] have been obtained from public internet sources. We also examined the statistics of the set of Twitter hashtags compiled by [18].

The algorithm used for the compilation of the Billboard charts has been adjusted over time, with a major update in 2014/15. At that point the traditional sales-based ranking was substituted by a ranking based on a multi-metric consumption rate, which includes weighted song streaming. This update, which took effect at the end of 2014, affects the chart statistics profoundly.

The original Billboard album chart, the Billboard Top 200, was retained after the 2014/15 metric update under a new name, as ‘Top Album Sales’. Data continuity is consequently achieved when using the Top Album Sales charts from 2014/15 on. Whenever possible, we show results obtained from both the Billboard Top 200 and the Top Album Sales charts, where the latter are compiled according to the unaltered sales-based ranking rules, albeit only for 100 ranks.

For data analysis, we used statistically relevant binning, which consists, as explained in Sect. 4.5, of adjusting bin sizes dynamically such that a minimal number of \(N_\mathrm{min}^\mathrm{(data)}\) data points per bin is obtained.

4.2 New York Times best-seller list

In contrast to the Billboard magazine, the New York Times does not publish a full history of their best-seller list. Only recent lists are published on the official New York Times website [26]. Consequently, the analysis presented here relies on a republication of the data by Hawes Publications [12], which collected and republished the weekly best-seller list from the 1950s until today. Since the length of the best-seller list changed repeatedly, all analysis above consider a top 10 ranking.

4.3 Reddit data analysis

Reddit data were downloaded from the Reddit Database [32]. Due to the large data size, we restricted our analysis to two six-month periods, the first half of 2013, which contains 19.050.122 post and 187.385.130 comments, and the second half of 2015, with a total of 37.038.895/347.236.797 entries. Comments of posts receiving less than 1000 comments over the observation period were removed from the data, since they would not show up in the charts analyzed. The cleaned comment data was separated into time slices based on the comment timestamps and aggregated by post IDs (identification labels). In this way charts for consecutive 10 min and 60 min slices where generated, as based on the number of comments received. Post lifetimes were evaluated from the respective top-100 charts.

4.4 Twitter hashtags

The hourly ranking of the 50 most common hashtags were gathered by Mønsted [24]. This dataset, which has been used to study the acceleration of collective attention [18], contains the 50 most common hashtags out of a 10% sample of all tweets, which were gathered every hour during the period from 2013 through 2016. For every hashtag the dataset also shows the number of times it appeared during the last hour in the 10% sample. This additional information allows, modulo the top-50 cut-off, to compile charts with longer observation intervals, e.g. daily or weekly charts.

4.5 Adaptive binning

For the generation of probability distribution functions from collected data one needs to group events together into bins. When sets of bins with predefined width are used, the number of events may vary strongly from one bin to another. It is hence advantageous, for statistical relevant binning, to adjust the width of the individual bins dynamically, until a minimum of \(N_\mathrm{min}^\mathrm{(data)}\) data points per bin has been reached. The binning procedure is finished once this is not any more possible. Comparing results obtained for different \(N_\mathrm{min}^\mathrm{(data)}\), as done in Fig. 5, allows to gauge the accuracy.

5 Discussion

Consumption charts, like sales and streaming charts, show remarkable similarities between different cultural products, e.g. when comparing music albums and literature. The evolution of the respective chart diversities, as well as the probability that a number-one title debuts as such, follow parallel courses for these two cultural goods. Similarly, one can fit the distribution of on-chart residence times, the respective lifetimes, in all cases by a generalized log-normal distribution that can be shown to maximize the information content of the lifetime distribution, to the extent as it is stored in the brain. Taking the long-term perspective, one notices distinct features for the evolution of the lifetime distribution of the New York Times best-seller list, and for album sales charts.

For music albums [34], the lifetime distribution evolves from a log-normal distribution, as before the 1990s, to a power law, as for nowadays. In this context it is important to notice that power laws belong to the class of generalized log-normal distributions, albeit with a vanishing quadratic term in the exponent. The information-theoretical arguments presented in this study for the occurrence of log-normal distributions suggests that a transition from fully log-normal to a power law takes place when uncertainties, viz the variance, fade into secondary importance. This phenomenon is related to the time needed on the average for decision-making, given that it takes longer to gauge variances than means.

In societies, some timescales remain constant, others are subject to a secular acceleration process [33]. An example of a non-changing time scale is the period of typically 4–5 years between general elections, which contrasts with an accelerating opinion dynamics [9]. Of relevance for the present study is the charting period of 1 week, which did not change since the inception of classical music charts and book bestseller lists. In contrast to the charting period, the decision time to buy an album is likely to have fallen substantially with the rise of the internet. Here we have argued that this development is reflected in the respective chart statistics. It is less clear if the same holds for the time people need to read a book.

Modern streaming charts can be used as a test bed for the relative time scale hypothesis. The available high temporal resolution allows narrowing down the time individuals need to decide to listen to an album, and then to actually do so. The lifetime distributions of Spotify album charts show that this time scale is nowadays between 1 week and 1 day. For a single song, the decision can be made comparatively faster, which leads for both weekly and daily single charts to power law lifetime distributions.

Activity charts, such as Twitter hashtags or Reddit postings, differ in certain aspects from consumption charts. Human activities are subject to a day–night cycle, which shows up for the case of commenting on Reddit as a resurgence after around 24 h. Previous-day posts are likely to be revisited after a night of sleep, as evident in Fig. 5. This phenomenon, the 24-h cycle, will mask underlying power laws, if existent. The Twitter charts provide evidence for an equivalent occlusion mechanism, see Fig. 3.

The lifetime distribution of hourly hashtag charts is characterized by a peak at 18 h, together with a subsequent drop at 24 h. Charts averaging over one or more full days, daily and weekly charts, are represented instead by smooth distributions. One finds indications for power laws with minor quadratic corrections. These results suggest that power laws appear ubiquitously for long-enough charting periods. In conclusion we believe that the data examined supports the basic hypothesis presented here, namely that aggregated human decision processes exhibit the pronounced statistical features characteristics of compressed information maximization.

Beyond information maximization, we found marked changes at the top of cultural charts, which started with the raise of the internet in the 1990s. This phenomenon regards the route to become a number-one hit, which took substantially longer in the past. We believe that this observation deserves further investigations, which would however transcend our present framework.