1 Introduction

Despite large advances in observation, numerical models and theory, the heating of the solar atmosphere remains an open question. The current question is perhaps not what the mechanisms for heating are, but rather which of the several proposed mechanisms dominate, and which different mechanisms are active in different regions. There are observational and theoretical reasons to support the bulk heating of plasma in the chromosphere, some proportion of which flows into the corona to maintain high temperatures, although the debate has not been settled (Aschwanden et al., 2008; Klimchuk and Bradshaw, 2014). The movement of kilogauss flux tubes creates stress and reconnection at small-scale current sheets – this is the nano-flare model of heating (Parker, 1988; Hansteen et al., 2015). Brightenings, or small flare-like events, can be observed in imaging and spectroscopic data. They occur at a huge range of energies and spatial scales, and large events occur less frequently than smaller flares. Fitting the energy distribution of observed small-scale brightenings to a power law provides an estimate of the energy available at unresolved scales (Crosby, Aschwanden, and Dennis, 1993; Hudson, 1991; Lu and Hamilton, 1991) and may give constraints on heating mechanisms. However, published results give a range of power laws which are inconclusive (Vilangot Nhalil et al., 2020; Aschwanden and Parnell, 2002).

Previous works have implemented a wide variety of methods for automatically detecting brightenings. Henriques et al. (2016) use binary mapping, detection thresholding and subsequently a detection filter based on characteristics such as length and duration. Sekse, Rouppe van der Voort, and De Pontieu (2012) attempt to determine whether events in neighbouring frames belong to the same chain of events based on a frame-by-frame pixel density overlap. Work has also been done with filtering long-time intensity trends from detections in order to determine the true brightness of white-light flares (Mravcová and Švanda, 2017), running central median methods for resolving fine-scale dynamics (Plowman, 2016), and the inference of background data and other properties from small boxes/cubes surrounding each brightening (Nelson et al., 2017a). There are few instances of comprehensive, homogeneous multi-instrument studies of small-scale structures/events conducted over very large data sets, that encompass and distinguish large regions of the chromosphere/corona. Some examples include Hou et al. (2016)’s spectral analysis of over 2700 AR Si iv dots, Nelson et al. (2013)’s large study of 3570 H\(\alpha \) Ellerman Bombs observations, and Tian et al. (2014)’s 176 multi-instrument penumbral dots study.

The goal of this study is to present an efficient brightening detection and characterising method, tested on synthetic data, that will be used for a large future study. Section 2 describes the filtering and data extraction method, using its application to synthetic data to aid the method description. Section 3 analyses the results of its application to synthetic data, and the optimisation of certain method parameters. The application of the method to IRIS observations of a quiet-sun (QS) region is also discussed, along with some preliminary results. A summary and description of future application of the method is given in Section 4.

2 Method

2.1 A Synthetic Datacube

A datacube of synthetic data is created in order to test the method, and to facilitate the description of the method. For this purpose, this datacube is deliberately kept simple, with a low number of bright points to avoid overlap, and with properties (brightness, area and duration) that facilitate detection. The background image for the synthetic datacube is created from a real IRIS slit-jaw observation time series. Details of the quiet-sun observations are given in Section 3.3. We take the mean intensity over time at each spatial pixel over the whole data set in order to create a background image, shown in Figure 1(a). This image is then replicated 120 times, forming a datacube over time of size \([352, 348, 120]\), a size typical of IRIS slit-jaw datacubes, and corresponding to the real datacube presented later. For each pixel, a random time series is created from a Poisson distribution based on that pixel’s mean intensity. Thus the synthetic data has a background intensity which varies randomly over time due to Poisson noise. Approximately 1285 brightenings are added to the datacube. These have Gaussian profiles of intensity in space and time, with random distributions of spatial and time position, area and duration, some of which can be seen in Figure 1(b). The spatial \(x\) and \(y\) Gaussian widths range from 0.5 to 3 pixels (standard deviations), and their durations vary from 1 to 4 time steps. These ranges are based on detected bright points in the real data. The Gaussian amplitudes (or brightness) are also set at random, with a distribution between 10 and 30 times the standard deviation of the datacube over time. Random linear motions are imposed to these brightenings ranging from 0 to 10 pixels over the bright point’s lifetime – loosely corresponding to the motions we see and detect in the real data. Figure 1(b) shows a frame from the datacube after the addition of Poisson noise and several Gaussian brightenings. These four components – a static smoothly-changing background, Poisson noise, random Gaussian brightenings and random linear motions – provide an appropriate analogue to real QS IRIS data.

Figure 1
figure 1

A simple representation of the creation of a synthetic datacube and the detection of a brightening therein. (a) shows a frame from the mean background of the synthetic data. (b) is the same as (a) but with Poisson noise and Gaussian brightenings added. (c) is image (b) following the filtering method. (d) is the same as (b) with a brightening highlighted by a red plus. The plus represents the position of maximum amplitude for this brightening.

2.2 Band-Pass Filtering

The first step of the detection method consists of a band-pass filter in space and time. The filtering is achieved by convolution with appropriate kernels along each dimension of the datacube. The kernels are formed using IDL’s DIGITAL_FILTER function (Walraven, 1984). Considerable efficiency can be achieved by applying convolution to the datacube \(S\) with vector kernels in the two spatial and one temporal dimension sequentially, rather than convolving once with a 3-dimensional kernel. The filtered cube \(F\) is defined as

$$ \begin{aligned} S^{\prime }&=S \ast \alpha (x), \\ S^{\prime \prime } &=S^{\prime }\ast \beta (y), \\ F & = S^{\prime \prime } \ast \gamma (t) , \end{aligned} $$
(1)

where \(\alpha (x)\), \(\beta (y)\) and \(\gamma (t)\) are the \(x\), \(y\) and \(t\) kernels, respectively. Note that in this work the \(x\) and \(y\) kernels are identical, although there may be instances where different band-pass parameters may be appropriate (e.g. for data close to the limb). We avoid this complication in this study by choosing data that is close to disk centre.

The most important parameters for these filter kernels are the low and high frequency values controlling the band-pass, \(f_{\mathit{low}}\) and \(f_{\mathit{high}}\). Increasing the \(f_{\mathit{low}}\) value reduces the power of slow temporal variation and spatially smoothly-changing features, whilst decreasing the \(f_{\mathit{high}}\) value reduces the power of high-frequency temporal variations (i.e. background noise). Figure 1(c) shows a frame of the datacube following filtering, with \(f_{\mathit{low}}=0.09\) and \(f_{\mathit{high}}=0.40\). We further illustrate the filtering effect in Figure 2 on a slice of the datacube signal which passes through a bright point. Three slices are shown, corresponding to cuts across the \(x\), \(y\), \(t\) dimensions. The unfiltered signal is shown as the black line. The red line shows the data following a convolution with a band-pass filter kernel with \(f_{\mathit{low}}=0.09\) and \(f_{\mathit{high}}=0.40\) (in units of fractions of the Nyquist frequency), showing effective dampening of the slowly changing background and high-frequency noise. The blue line shows a varying threshold parameter which is described in the following subsection: any red points above the blue threshold are detected as a candidate brightening event. The selected brightening event is made prominent by the filtering, and is clearly above the threshold. The choice of the \(f_{\mathit{low}}\) and \(f_{\mathit{high}}\) frequencies affects the results, and will be discussed later in the context of optimising the method’s performance on the synthetic datacube.

Figure 2
figure 2

A demonstration of the filtering and thresholding method, showing slices through the datacube passing through a bright point across the \(x\) (left), \(y\) (centre), and time dimensions (right). Unfiltered synthetic data is shown in black. Filtered results are shown in red, and the blue line denotes a threshold value that enables detection of the narrow synthetic brightenings (see text). The horizontal dashed line shows the position of the maximum brightness of the bright point.

2.3 Thresholding for Initial Detection

We estimate a threshold which is based on the Poisson noise model for each spatial pixel, thus the threshold varies over the field of view, but remains constant over time. Whilst the Poisson noise level is equal to the square root of the mean signal over time, we wish to define a threshold based on the filtered, rather than original, data. We therefore calculate a threshold numerically. A datacube is generated in the same manner as the synthetic data above, but without the addition of bright points – so a static background with time-varying Poisson noise. This cube is then band-pass-filtered over all dimensions as described above. A spatially-varying threshold (\(T_{\mathit{high}}\)) is then defined as the standard deviation over time, \(\sigma \), of the filtered datacube multiplied by a constant. This constant is discussed in the following section.

The red cross in Figure 1(d) indicates a particular region from Figure 1(c) that is above this threshold. This is considered as a detection, or a candidate brightening event. This event corresponds to the bright point shown in Figure 2. Following thresholding, candidate brightenings that are smaller than 25 voxels or last less than 5 frames are discarded. For each detected event, we take the extended surrounding region and apply a lower threshold (\(T_{\mathit{low}}\)) such that any voxels within this local region that are above \(T_{\mathit{low}}\) and which contain the original bright points detected using \(T_{\mathit{high}}\) are defined as the final detected bright point: thus the small region identified using \(T_{\mathit{high}}\) is grown to a larger region using \(T_{\mathit{low}}\). This two-step process using an initial high threshold then a lower threshold enables effective isolation of separate bright points and increases the number of voxels which more accurately matches the total brightness of that bright point.

For each detected region, the following information is recorded:

  • The duration of the event

  • The volume, defined as the number of all voxels contained in a brightening event

  • The background brightness, defined as the median brightness of the voxels of the extended region surrounding the bright point which are below \(T_{\mathit{low}}\)

  • The maximum brightness, or the count at the brightest voxel minus the estimated background intensity

  • Total brightness, or the counts minus the background count summed across all the bright point voxels

  • The maximum brightness at each time step

  • The total brightness at each time step

  • The spatial centroid of each region during each time frame, weighted by the counts at each voxel

  • \(N_{\mathit{frag}}\); many detected brightenings in real data do not maintain a single, coherent region over time – single regions can fragment into two or more regions, or several isolated regions may merge. This is either real, a consequent of noise, is due to some parts of a brightening becoming faint and falling below the detection threshold, or is due to the complexity of the underlying ‘static’ structure affecting the detection method. For each brightening event, the number of fragments per time step is recorded, and the maximum number of fragments over the event’s lifetime is recorded as value \(N_{\mathit{frag}}\). For example, if an event remains unfragmented, then \(N_{\mathit{frag}}=1\) or if an event fragments to two regions at least once during its lifetime, then \(N_{\mathit{frag}}=2\), and so on. This is an approximate measure of how spatially coherent a brightening is over its lifetime.

  • Average speed; for events with \(N_{\mathit{frag}}=1\), the speed between time steps is calculated from the spatial centroid and the observational cadence, and an average speed calculated over the event’s lifetime. A more sophisticated analysis is required for \(N_{\mathit{frag}}>1\), reserved for future work.

For application to real data, \(N_{\mathit{frag}}\) is useful to quickly isolate the most coherent brightenings for further study. This characteristic will be explored further in the Results & Discussion section. We initially explored the possibility of fitting these bright point profiles to Gaussian functions in space and time. This was unsuccessful due to noise, the complexity of the bright points (particularly considering fragmentation), the high numbers of parameters needed to fully describe a moving Gaussian shape in three dimensions, and the small number of detected voxels for some candidate events. Furthermore, Vissers et al. (2015) suggest that a Gaussian model may not be appropriate for all transition region (TR) phenomena.

3 Results & Discussion

3.1 Synthetic Datacube

We apply the method to the synthetic datacube. The results are compared to the known distribution of the synthetic brightenings as a test of the method, and used to optimise various parameters. There are two ways to assess our method using synthetic data: a voxel-by-voxel set of statistics, and an event-by-event comparison. For the voxel-by-voxel statistics, we calculate the following:

  • True Positive (TP) – the number of detected voxels which correspond to synthetic brightening voxels

  • False Positive (FP) – the number of detected voxels which do not correspond to synthetic brightening voxels

  • False Negative (FN) – the number of non-detected voxels that correspond to synthetic brightening voxels

  • True Negative (TN) – the number of non-detected voxels that correspond to voxels without synthetic brightenings

Note that we define a synthetic brightening voxel as voxels that have a brightness greater than 30% of their respective brightening’s maximum amplitude, calculated from each brightening’s Gaussian parameters. These criteria also define the Precision and Accuracy percentages, which are determined as:

$$\begin{aligned} \mathrm{Precision}&=\frac{\mathit{TP}}{\mathit{TP}+\mathit{FP}}\times 100 \%, \end{aligned}$$
(2)
$$\begin{aligned} \mathrm{Accuracy}&=\frac{\mathit{TP}+\mathit{TN}}{\mathit{TP}+\mathit{TN}+\mathit{FP}+\mathit{FN}}\times 100 \%. \end{aligned}$$
(3)

For an event-by-event comparison, a simple criterion we use is the distance of a detection’s centroid to the closest true brightening, \(\Delta =\sqrt{\Delta x^{2}+\Delta y^{2}+\Delta t^{2}}\). This criterion allows a percentage scoring of the method with regards to accurate event detection, defined as

$$ \frac{n_{\mathit{good}}}{n_{\mathit{total}}}\times 100 \%, $$
(4)

where \(n_{\mathit{good}}\) is the number of detections with the stringent criterion \(\Delta \leq 10\), and \(n_{\operatorname{total}}\) is the total number of brightenings in the synthetic datacube. Detections with \(\Delta >10\) are considered spurious.

Table 1 shows various values of these criteria based on different values of the \(f_{\mathit{low}}\), \(f_{\mathit{high}}\), \(T_{\mathit{low}}\) and \(T_{\mathit{high}}\) parameters. For the synthetic data, we find an optimal band-pass at \(f_{\mathit{low}}=0.09\), \(f_{\mathit{high}}=0.40\) with threshold values of \(T_{\mathit{low}}=7\) and \(T_{\mathit{high}}=9\), whereby the detection percentage mentioned above reaches \(\sim 81\%\) and the TP detection reaches \(\sim 50\%\) while FPs remain \(<10\%\) and event-by-event accuracy remains high. The rest of this work is based on these parameters.

Table 1 Comparison of some frequency limits, threshold values and their effect on detection precision. \(f_{\mathit{low}}=0.09\), \(f_{\mathit{high}}=0.40\), \(T_{\mathit{low}}=7\) and \(T_{\mathit{high}}=9\) are chosen as the parameters for use with real IRIS data.

For ‘true’ event detections (\(\Delta \leq 10\)), Figure 3(a) shows that there is a good correspondence between the detected maximum brightness and the true maximum brightness, with a mean absolute relative deviation (MARD) of 10.32% and a gradient of 1.13. Figure 3(b) compares the detected and true speed of the bright point, with a gradient of 1.04 and a MARD of 130.81%. Figure 3(c) shows the detected and true total brightness with a gradient of 0.32 and MARD \(59.60\%\). This gradient of 0.32 shows that the total brightness is greatly underestimated. Since the maximum brightness of an event is accurately found, this underestimation is due to non-detection of brightening voxels, or an underestimation of a brightening’s volume. This can be improved by decreasing \(T_{\mathit{low}}\), but at the expense of increasing false positives – our choice of parameters is therefore a compromise. Figure 3(d) shows a histogram of \(\Delta \) values, \(\sim 99\%\) of which are \(\le 3\). These comparisons give approximate estimates of the uncertainty levels which we can assign to the brightnesses of detections in real data: the uncertainty of maximum brightness amplitude is small compared to general radiometric calibration errors, whilst the uncertainty in the total brightness is large, and comparable to the calibration errors.

Figure 3
figure 3

Comparisons of detected and true properties of the synthetic bright points: (a) synthetic amplitudes vs. detected brightness; (b) synthetic speed vs. detected speed; (c) synthetic total brightness vs. detected total brightness; and (d) a histogram of \(\Delta \) values for the event-by-event analysis.

All synthetic detections are dominated by \(N_{\mathit{frag}}=1\) values, with some detections with \(N_{\mathit{frag}}=2\). In an analysis of real data, a user can choose \(N_{\mathit{frag}}=1\) values only if they want to exclude overlapping or dubious detections. However, as stated in Section 2, analysing the distribution of \(N_{\mathit{frag}}\) gives an idea of the complexity of these events.

3.2 A More Complicated Data Set

A second synthetic datacube is created containing \(\approx 10^{4}\) brightenings. The area, duration and maximum brightness of these brightenings follow continuous power-law distributions. The lower limits of these distributions are set at values below that which we expect to detect. The lower limit for the maximum brightness amplitude of the Gaussian events is 10 DN which is equal to the minimum \(2\sigma \) noise level of the background. The lower limit for the spatial \(1\sigma \) width is 0.4 pixels, and the lower limit for the temporal \(1\sigma \) width is 0.25 time steps. The slopes of the power law distributions, as set by the standard parameter \(\alpha \) (see e.g. Clauset, Shalizi, and Newman, 2007), are \(\alpha _{B} = 2\), \(\alpha _{A} = 3\), and \(\alpha _{t} = 2.5\), for brightness, area and duration, respectively. These \(\alpha \) values are set so that the largest values of each property are at expected values. That is, so the events do not become too large or long-lived. These power laws are shown as the black lines in Figure 4. We assume that there is an approximate linear relationship between brightness, area and duration. For example, we expect the brightest events to possess, in general, a larger area and duration. We use a local randomisation process to impose these approximate relationship from the initial distributions, with the relationships shown in the plots of Figure 5. The events have a uniform distribution of speeds ranging from zero to two pixels per time steps, in random directions, and are placed in random positions throughout the datacube. Due to their speed and position, some events may appear or disappear at the datacube edges over time, and many events may overlap. In the case of overlaps, the maximum brightening at a given pixel is retained.

Figure 4
figure 4

Input (black) and detected (red) distributions of (a) maximum brightness, (b) total brightness, (c) duration (measured in time steps or frames), and (d) spatial area (measured in pixels), for the 10 000-event synthetic datacube. Axes are \(\log _{10}\) of each characteristic.

Figure 5
figure 5

The relationship between input values for the 10 000 brightenings, showing (a) the \(1\sigma \) Gaussian area against the \(1\sigma \) Gaussian time (or duration), and (b) the maximum brightness (Gaussian peak amplitude above the background) against duration.

This datacube is likely more representative of real-world IRIS event distributions and is a more realistic test for the detection method, whereby a large number of events are difficult or impossible to distinguish from noise, and many other events will not meet the filtering criteria, i.e. some events will be too small or large, too faint, or too short-/long-lived. The same filtering process and parameters described in the previous sections are applied to this datacube. About \(23\%\) of the 10 000 events are detected. For these detections, 88% are recorded as one coherent event (\(N_{\mathit{frag}}=1\)), and 9% split into two regions at least once during their detected lifetimes (\(N_{\mathit{frag}}=2\)).

The resulting detected power laws are shown in red in Figure 4. Figure 4(a) shows a reasonable agreement between the slopes of the input/detected maximum brightness distributions only above \(\approx 160\) DN. Below this limit, many events are detected, but at a small fraction of the true number. This fraction decreases rapidly with decreasing brightness. Above this limit, the number of detected events is slightly larger than the true number, showing an overestimation of brightness for some events. This is likely due to overlaps between brightenings, or an underestimation of the background brightness. A similar distribution holds for the total brightness. In analysing distributions of detections in real data, this decrease in detection rate for dimmer events must be properly modelled, for example following the approach of Clauset, Shalizi, and Newman (2007). Without accounting for the efficiency of the method to various brightness values, the power law slope \(\alpha \) will always be greatly underestimated. This will be explored further in future work. Although detections may not directly correspond to input synthetic brightenings, the distribution of the detections still depends on the distribution of the synthetic inputs, even if the relationship becomes nonlinear. The nature of the detection process fundamentally changes the distribution of detections in a stochastic fashion and therefore a one-to-one relationship between true events and those characterised by detection algorithms is not always possible. Additionally, this study is limited to the 1400 Å channel. It is possible, and indeed likely, that these brightening events could be emitting in several other wavelengths, which would suggest that energy output measurement are systematically underestimated by using only one channel. Indeed, some events may only exist at higher temperatures and energies, and this will have a direct impact on power-law gradients. This is similar to the arguments given by Aschwanden and Parnell (2002) and Aschwanden and Charbonneau (2002) in the context of nanoflares at much higher temperature ranges. The distribution of detected durations shown in Figure 4(c) follows the distribution of input durations well for events with \(\sigma \) time width of \(\approx 1\) or longer. Some durations are overestimated – this is likely due to overlaps between brightenings. No events with \(\sigma \) time widths of less than 0.6 are detected – this is due to the method discarding detections below this limit in order to reduce the number of false positives. The area distribution of Figure 4(d) follows the input distribution well above a limit of \(\approx 0.9\) pixel \(\sigma \) area, with some event areas overestimated. The method does detect many small-area events below this limit, but at a fraction that decreases with decreasing area. While the detected distribution matches the input distribution well down to a lower limit, this does not necessarily imply a one-to-one correspondence between a match and its true value. The slope of each distribution does not depend on the exact value of any detection (due to the scale-invariant nature of power-laws); only how they are distributed.

Figure 6 shows a one-to-one comparison of input and detected events. The distance (or \(\Delta \), as defined previously) between all input events and a detection is calculated, and the input event with the smallest \(\Delta \) is paired to that detection. For comparison, we only consider the \(\sim 11\%\) of input events with \(\Delta \leq 3\). The distribution of \(\Delta \) for these selected pairs is shown in Figure 6(a). To statistically compare the detected and input properties, we fit the scattered points to a straight line using a least-absolute-deviation (LAD) fitting procedure, and calculate the median relative deviation (MedRD, which will be positive/negative for systematic over/underestimation) and the median absolute relative deviation (MedRAD, which is the spread of deviations from the MedRD). The LAD fitted lines are shown in black in Figures 6(b)–(f). Note that the fitting is done in log space for Figures 6(c)–(d), and in linear space otherwise. The blue lines in Figures 6(b)–(f) indicate the exact one-to-one relationship.

Figure 6
figure 6

Plots comparing input and detected properties for the \(\approx 1100\) paired input–output events (see text): (a) distribution of \(\Delta \) values (or spatio-temporal distance), (b) detected vs. true average area, (c) detected vs. true maximum brightness, (d) detected vs. true total brightness, (e) detected vs. true duration, and (f) detected vs. true speed. The blue lines indicate one-to-one relationships while the black line is a least-absolute-deviation (LAD) fit to the points. Note that the LAD fit is done in log-space for the brightness of (c) and (d), with the others fitted in linear space. The black and blue lines in (c) overlap.

Figure 6(b) compares the input and detected areas for the paired brightenings. The detected areas tend to underestimate the true areas with a MedRD of −15%, with an increasing underestimation with increasing area shown by the LAD fit gradient of 0.5. The scatter of points is high, with a MedRAD of 30%. Figure 6(c) shows that the detected maximum log brightness gives a good estimate of the true maximum log brightness for the bulk of events that obey the \(\Delta \leq 3\) criterion (with a gradient of \(\sim 1\) whereby the dashed black and blue lines overlap and a MedRD close to zero), with large scatter only for very faint events, and a MedRAD of 6%. However, across the detections as a whole, without applying this \(\Delta \) proximity criterion, the correspondence between inferred and input maximum brightness drops off steeply for dimmer events, as can be seen in Figure 4(a). The results of Figure 4(a)’s histogram begin to match with those of Figure 6(c) only for the brightest events. The true log brightness shown in Figure 6(d) shows an overwhelming underestimation of log-brightness for the bulk of events with a MedRD value of \(-30\%\). The gradient of 1.1 shows that the larger/brighter events give a better estimate of total brightness, with a large scatter for dim/small events and an overall MedRAD of 22%. The duration results shown in Figure 6(e) are similar to the area results from Figure 6(b) – there is a tendency to underestimate (MedRD of \(-15\%\)), with the underestimation increasing with increasing duration (LAD gradient of 0.5), and a large scatter (MedRAD of 25%). Figure 6(f) shows the performance of the method in tracking the centroid of events over time through comparing average speeds. The comparison is reasonable given the challenging task. Underestimations of true speed values (i.e. detected values of speed) become more severe as true speed increase, with a gradient of 0.6, a MedRD of \(-16\%\) and a MedRAD of 23%.

These comparisons will be central in interpreting the results, and power laws in particular, from real data in future studies, and highlight the difficulty of analysing power laws of small-scale brightenings in solar data. The one-to-one comparisons (only possible, of course, on synthetic data) show that only the detected maximum brightness can be considered a reliable measure for the brightest and most accurately detected events with a low uncertainty of \(\approx 6\%\). With real data, one way of confirming whether a detection is true or not is to calculate \(\Delta \) across several wavelength channels. If an event appears across several channels and these facets are within an acceptable proximity to each other then the event is deemed to be true and the maximum brightness values recorded for those events are reliable. This will be implemented in a subsequent paper. The other values are mainly useful for showing trends or relationships for large statistical samples, with a systematic underestimation of area, duration, and total brightness, and a general statistical uncertainty of around 30%.

3.3 IRIS Data

We analyse one data set from IRIS (De Pontieu et al., 2014) centred at approximately X = \(-126.05''\), Y = \(5.87''\). The observations begun on October 5, 2013 at 08:17 UT and ended the same day at 08:36. Figure 7 (left) shows the region of interest (ROI) as a QS region. The IRIS data consist of a 1400 Å  sit-and-stare series of slit-jaw images, with a consistent 9.46 s temporal cadence, 0.17″ pixel−1 spatial scale and a \(57.89''\times 58.56''\) field of view (FOV).

Figure 7
figure 7

IRIS 1400 Å slit-jaw images of (left) the ROI with (right) an over-plot of all detections from the same time frame. The left image has been processed such that isolated pixels or small regions of outlying intensity (such as the central slit) are treated as missing data. An animation of this figure is available. The video begins on Oct. 5th 2013 at 08:17 UT. The video ends the same day at 08:36.

The IRIS data set is processed using standard procedures to account for temporal exposure, and empty margins are removed. Dark current subtraction as well as flat field, geometrical and orbital variation corrections have been applied as standard procedure for level 2 IRIS data. Isolated pixels with spuriously high values are flagged as missing data, and are ignored in the filtering and detection method. The central slit and its surrounding pixels are also treated as missing data such that extremely ‘dark’ pixels do not affect the filtering process. Additionally, drift due to solar rotation has been treated using a Fourier Local Correlation Tracking method (Fisher and Welsch, 2008).

The results of the detection method with the IRIS datacube can be seen in Figure 7, whereby the left plot represents the first image of the IRIS datacube. The right plot is the same IRIS image with all brightenings detected in the same time frame as the image on the left. Yellow contours represent brightenings with \(N_{\mathit{frag}}\le 2\), green represents \(N_{\mathit{frag}}=3\) or \(N_{\mathit{frag}}=4\), and blue represents \(N_{\mathit{frag}}\ge 5\). 2997 small-scale brightening events are detected within the \(54.40''\times 55.23''\) FOV and 18.76-minute observation period, whereby 549 detections correspond to \(N_{\mathit{frag}}=1\) and 791 detections correspond to \(N_{\mathit{frag}}=2\), as can be seen in Figure 8. This yields a combined event density of \(3.96\times 10^{-4}\) events per second per arcsec2. The remainder consist of \(N_{\mathit{frag}}\) values between 3 and 12. This distribution of fragmented detections is quite different to the test datacube of Section 3.2, and shows that the brightening detections in IRIS data tend to be more incoherent and fragmented. A more detailed analysis can show whether this is due to a true property of brightenings, or due to some other effect (e.g. high time variability of the background brightness, on which brightening events are superimposed, or overlapping of events in high-activity regions). For the preliminary results presented here, we have chosen to only include brightenings with \(N_{\mathit{frag}}=1\).

Figure 8
figure 8

Histogram of fragmentation during the lifetime of all detections: 549 events remain un-fragmented with \(N_{\mathit{frag}}=1\), 791 detections fragment once with \(N_{\mathit{frag}}=2\), 658 events have \(N_{\mathit{frag}}=3\), 414 events have \(N_{\mathit{frag}}=4\), and the remaining 585 detections have \(N_{\mathit{frag}}\) values between 3 and 12.

The choice of \(f_{\mathit{low}}\) and \(f_{\mathit{high}}\) has substantially reduced both small-scale, rapid noise and large-scale, slow changes, as can be seen in Figure 9. The vertical dashed lines represent the coordinates of a real detection according to its maximum brightness values. It is clear from these plots, similar plots made for other detection events, as well as values of \(\Delta \) for synthetic data, that the method is effective at identifying and accurately determining the location of detections.

Figure 9
figure 9

An example of the filtering process on slices taken through the IRIS observation datacube’s (left) spatial \(x\), (centre) spatial \(y\), and (right) temporal \(t\) dimensions, passing through a point coinciding with an event detection, denoted by the vertical dashed line. The black line is the level-2 IRIS signal, the blue denotes the threshold, and red is the filtered data. Intensity values are measured in DN s−1.

Figure 10 presents some preliminary statistical results of these detections’s properties: average speed (left), spatial area (centre) and total brightness (right). Most detections have a speed between 3 and 15 km s−1 (with a mean value of 9.03 km s−1), cover an approximate average area between 0.2 and 0.4 arcsec2 (with a mean value of 0.32 arcsec2) and have a total brightness of 1600–31600 DN (with a mean value of 6300 DN). Large and/or bright events are rarer than small and/or dim events, and the distributions are all approximately symmetrical. As mentioned in Section 3.2, the detection process is likely limited by observational resolution which would explain the distribution drop-off for smaller/dimmer/slower events. Figure 10 (right)’s profile, above the mean log-brightness of \(\sim 3.8\), suggests that these brightenings follow a power-law distribution. Distributions such as these will be studied in more detail in future work.

Figure 10
figure 10

Basic properties of the detections from the Oct. 5th 2013 IRIS data. The left, centre and right panels plot histograms of average speed, average area and total brightness, respectively. The dotted vertical lines represent the mean of each property; 9.03 km s−1, 0.32 arcsec2 and 6300 DN s−1 for speed, area and total brightness, respectively.

Figure 11 shows log–log scatter plots of area vs. duration (a), total brightness vs. duration (b), total brightness vs. area (c) and volume vs. maximum brightness (d). The green line shows the result of a robust linear fit to the logarithmic values, giving gradients of 0.14 (a), 1.60 (b), 1.32 (c), and 0.45 (d). The red triangle represents the mean value of both variables within each plot. The shallow gradient in Figure 11(a) between area and duration suggests that there is no relationship between these properties. However, this result may be limited by the IRIS temporal cadence; the detection process detects the duration of events in integer number of frames, which can be seen by the sudden jumps in values along the y-axis. The plots of Figures 11(b)–(d) show a clear proportional relationship between duration with total brightness, area with total brightness, and volume with maximum brightness, respectively. The former two relationships are unsurprising considering that a measurement of total brightness is dependent on the number of voxels. The latter relationship suggests that larger events tend to be intrinsically brighter than smaller events.

Figure 11
figure 11

Log–log scatter plots showing relationships between (a) area and duration, (b) duration and total brightness, (c) area and total brightness, and (d) volume vs. maximum brightness. The green lines represent a robust linear fit to all points. The red triangle shows the mean value of both variables.

Figure 12 shows the power of the method in analysing the properties of a brightening over its lifetime. Figure 12(a) tracks the motion of the brightening, which has an average speed of 0.06 pixel s−1. Figure 12(b) shows the change in total brightness over time (measured in DN). Figure 12(c) shows the change in area over time. Figure 12(d) shows the change of maximum brightness over time. We find that this profile is typical of a large number of brightenings which have the property \(N_{\mathit{frag}}=1\), with a coherent motion and smooth increases/decreases in area and brightness during the event’s lifetime. While increases in brightness and area intuitively coincide, Figure 12(d) suggests that the intrinsic brightness also increases and decreases (independent of the size of the event).

Figure 12
figure 12

Moment-by-moment properties of a single brightening event, including (a) the movement of the brightening during its lifetime (from black to red), (b) the change in total brightness over time (measured in DN), (c) the change in area over time (measured in arcsec2), and (d) the change of total intrinsic brightness per pixel (measured in DN).

A more detailed analysis of these relationships, applied to this and other data sets, will be made in a future study. We can draw comparisons between these and other studies of small-scale brightening phenomena, such as: H\(\alpha \) Ellerman Bombs (Ellerman, 1917), which demonstrate typical widths, lengths and durations of \(0.63''\), \(0.35''\) and 120 s, respectively (Nelson et al., 2017b); UV IRIS bombs of approximately \(2''\times 2''\) area (Chen et al., 2019); and CA ii chromospheric brightenings lasting at least 1.5 minutes over a \(2''\times 2''\) area (Kuckein et al., 2017). Additionally, Figure 10, (left) and (centre), demonstrates similar distributions to those of Hou et al. (2016)’s smaller area results and Tian et al. (2014)’s shorter duration results, respectively. Perhaps the results of this paper describe Vissers et al. (2015)’s Flaring Arch Filaments rather than larger and longer-duration phenomena, which may explain the motion of some of these brightenings. We note that most of these previous studies focus on spectroscopic analyses which will lead to differences between their results and those of this study.

4 Conclusions and Future Work

We present a general method to detect small-scale brightenings in solar time-series imagery. The method is applied to IRIS data sets in the UV, but can be applied to any time-series image data given appropriate adjustment of a few parameters. For a simple test data set of IRIS-like synthetic slit-jaw image time series with randomly-distributed, randomly-moving small-scale Gaussian brightenings, the filtering and detection method yields an event-by-event detection rate of \(\sim 81\%\) and a voxel-by-voxel detection rate of \(\sim 49\%\) (with a voxel-by-voxel FP rate of \(<10\%\)). The detected maximum brightness values provide good accuracy to the true values for the brightest and most accurately detected events. The speeds are also reliable, albeit with a higher scatter. Care must be taken with measurements of total brightness – our tests show a large underestimation (0.3 on average) of total brightness due to the number of fainter voxels not being detected. The method is also applied to a second datacube with a high density of events possessing a power law distribution of brightness, area and duration. This test demonstrates the difficulty of extracting power law properties from real data due to instrument and method sensitivities to a limited range of brightness, duration and area. However, even though detected properties can have a large scatter from the true values, the shape of the detection histograms largely match the original distributions over a certain range. It also shows that the method gives an estimate of event maximum brightness with low uncertainty. Other properties (area, duration, speed, total brightness) have a large uncertainty, but the method gives approximate proportional relationships over large statistical samples.

In total 2997 brightenings are detected in real IRIS data within a \(54.40''\times 55.23''\) area spanning 19 minutes; 1340 of these detected events either remain unfragmented or fragment at least once over their lifetime. These detections yield an approximate event density of \(3.96\times 10^{-4}\) arcsec−2 s−1. Most detections have an average speed between 3 and 15 km s−1 (with a mean value of 9.03 km s−1), cover an approximate area between 0.2 and 0.4 arcsec2 (with a mean value of 0.32 arcsec2) and emit a total brightness of 1600–31600 DN (with a mean value of 6300 DN). Linear fitting of brightening properties shows that there are proportional relationships between the brightenings’ area with total brightness, duration with total brightness, and volume with maximum brightness. However, there seems to be little proportionality between the brightenings’ area with duration: this may be limited by IRIS’s temporal cadence. Preliminary frame-by-frame plotting of a brightening event demonstrates coherent motion and a change in total brightness, area and intrinsic brightness over its lifetime – this seems typical of a large number of brightenings.

Having demonstrated the method’s performance, we plan to apply it to a larger sample of quiet-sun IRIS data. This future study will use all the IRIS channels, as well as co-aligned data from other EUV imaging instruments, allowing a statistical study of brightening properties across a range of temperatures/heights. We also plan to test the method on ground-based photospheric data taken in visible wavelengths. The Interactive Data Language (IDL) software described in this paper will be released to the community in the near future through the SolarSoft library.