Detecting and characterising small-scale brightenings in solar imaging data

Observations of small-scale brightenings in the low solar atmosphere can provide valuable constraints on possible heating/heat-transport mechanisms. We present a method for the detection and analysis of brightenings and demonstrate its application to IRIS EUV time-series imagery. The method uses band-pass filtering, adaptive thresholding and centroid tracking, and records an event's position, duration, and total/maximum brightness. Area, brightness, and position are also recorded as functions of time throughout their lifetime. Detected brightenings can fragment or merge over time, thus the number of distinct regions constituting a brightening event is recorded over time and the maximum number of regions are recorded as a simple measure of an event's coherence/complexity. A test is made on a synthetic datacube composed of a static background based on IRIS data, Poisson noise and $\approx10^4$ randomly-distributed, moving, small-scale Gaussian brightenings. Maximum brightness, total brightness, area, and duration follow power-law distributions and the results show the range over which the method can extract information. The recorded maximum brightness is a reliable measure for the brightest and most accurately detected events with an error of 6%. Area, duration, and speed are generally underestimated by 15% with an uncertainty of 20-30%. Total brightness is underestimated by 30% with an uncertainty of 30%. Applying this method to real IRIS QS data spanning 19 minutes over a 54.40"$\times$55.23"FOV yields 2997 detections. 1340 of these either remain un-fragmented or fragment to two distinct regions at least once during their lifetime equating to an event density of $3.96\times10^{-4}$arcsec$^{-2}$s$^{-1}$. The method will be used for a future large-scale statistical analysis of several QS data sets from IRIS, other EUV imagers, as well as H-$\alpha$ and visible photospheric imagery.

A datacube of synthetic data is created in order to test the method, and to facilitate the description of the method. For this purpose, this datacube is deliberately kept simple, with a low number of bright points to avoid overlap, and with properties (brightness, area, and duration) that facilitate detection. The background image for the synthetic datacube is created from a real IRIS slit-jaw observation time series. Details of the quiet-sun observations are given in section 3.3. We take the mean intensity over time at each spatial pixel over the whole data set in order to create a background image, shown in figure 1a. This image is then replicated 120 times, forming a datacube over time of size [352,348,120], a size typical of IRIS slit-jaw datacubes, and corresponding to the real datacube presented later. For each pixel, a random time series is created from a Poisson distribution based on that pixel's mean intensity. Thus the synthetic data has a background intensity which varies randomly over time due to Poisson noise. ∼ 1285 brightenings are added to the datacube. These have Gaussian profiles of intensity in space and time, with random distributions of spatial and time position, area, and duration, some of which can be seen in figure 1b. The spatial x and y Gaussian widths range from 0.5 to 3 pixels (standard deviations), and their durations vary from 1 to 4 time steps. These ranges are based on detected bright points in the real data. The Gaussian amplitudes (or brightness) are also set at random, with a distribution between 10 and 30 times the standard deviation of the datacube over time. Random linear motions are imposed to these brightenings ranging from 0 to 10 pixels over the bright point's lifetime -loosely corresponding to the motions we see and detect in the real data. Figure 1b shows a frame from the datacube after the addition of Poisson noise and several Gaussian brightenings. These four components -a static smoothly-changing background, Poisson noise, random Gaussian brightenings and random linear motions -provide an appropriate analogue to real QS IRIS data.

Band-pass filtering
The first step of the detection method consists of a band-pass filter in space and time. The filtering is achieved by convolution with appropriate kernels along each dimension of the datacube. The kernels are formed using IDL's DIGITAL FILTER function (Walraven 1984). Considerable efficiency can be achieved by applying convolution to the datacube S with vector kernels in the two spatial and one temporal dimension sequentially, rather than convolving once with a 3-dimensional kernel. The filtered cube F is defined as where α(x), β(y) and γ(t) are the x, y and t kernels respectively. Note that in this work the x and y kernels are identical, although there may be instances where different band-pass parameters may be appropriate (e.g. for data close to the limb). We avoid this complication in this study by choosing data that is close to disk center. The most important parameters for these filter kernels are the low and high frequency values controlling the band-pass, f low and f high . Increasing the f low value reduces the power of slow temporal variation and spatially smoothly-changing features, whilst decreasing the f high value reduces the power of high-frequency temporal variations (i.e. background noise). Figure 1c shows a frame of the datacube following filtering, with f low = 0.09 and f high = 0.40. We further illustrate the filtering effect in figure 2 on a slice of the datacube signal which passes through a bright point. Three slices are shown, corresponding to cuts across the x, y, t dimensions. The unfiltered signal is shown as the black line. The red line shows the data following a convolution with a band-pass filter kernel with f low = 0.09 and f high = 0.40 (in units of fractions of the Nyquist frequency), showing effective dampening of the slowly changing background and high-frequency noise. The blue line shows a varying threshold parameter which is described in the following subsection: any red points above the blue threshold are detected as a candidate brightening event. The selected brightening event is made prominent by the filtering, and is clearly above the threshold. The choice of the f low and f high frequencies affects the results, and will be discussed later in the context of optimising the method's performance on the synthetic datacube. We estimate a threshold which is based on the Poisson noise model for each spatial pixel, thus the threshold varies over the field of view, but remains constant over time. Whilst the Poisson noise level is equal to the square root of the mean signal over time, we wish to define a threshold based on the filtered, rather than original, data. We therefore calculate a threshold numerically. A datacube is generated in the same manner as the synthetic data above, but without the addition of bright points -so a static background with time-varying Poisson noise. This cube is then band-pass-filtered over all dimensions as described above. A spatially-varying threshold (T high ) is then defined as the standard deviation over time, σ, of the filtered datacube multiplied by a constant. This constant is discussed in the following section.
The red cross in figure 1d indicates a particular region from 1c that is above this threshold. This is considered as a detection, or a candidate brightening event. This event corresponds to the bright point shown in figure 2. Following thresholding, candidate brightenings that are smaller than 25 voxels or last less than 5 frames are discarded. For each detected event, we take the extended surrounding region and apply a lower threshold -T low -such that any voxels within this local region that are above T low and which contain the original bright points detected using T high , are defined as the final detected bright point: thus the small region identified using T high is grown to a larger region using T low . This two-step process using an initial high threshold then a lower threshold enables effective isolation of separate bright points and increases the number of voxels which more accurately matches the total brightness of that bright point.
For each detected region, the following information is recorded: • The duration of the event.
• The volume, defined as the number of all voxels contained in a brightening event.
• The background brightness, defined as the median brightness of the voxels of the extended region surrounding the bright point which are below T low .
• The maximum brightness, or the count at the brightest voxel minus the estimated background intensity. • Total brightness, or the counts minus the background count summed across all the bright point voxels.
• The maximum brightness at each time step.
• The total brightness at each time step.
• The spatial centroid of each region during each time frame, weighted by the counts at each voxel.
• N f rag . Many detected brightenings in real data do not maintain a single, coherent region over time -single regions can fragment into two or more regions, or several isolated regions may merge. This is either real, a consequent of noise, is due to some parts of a brightening becoming faint and falling below the detection threshold, or is due to the complexity of the underlying 'static' structure affecting the detection method. For each brightening event, the number of fragments per time step is recorded, and the maximum number of fragments over the event's lifetime is recorded as value N f rag . For example, if an event remains unfragmented, then N f rag = 1 or if an event fragments to two regions at least once during its lifetime, then N f rag = 2, and so on. This is an approximate measure of how spatially coherent a brightening is over its lifetime.
• Average speed. For events with N f rag = 1, the speed between timesteps is calculated from the spatial centroid and the observational cadence, and an average speed calculated over the event's lifetime. A more sophisticated analysis is required for N f rag > 1, reserved for future work.
For application to real data, N f rag is useful to quickly isolate the most coherent brightenings for further study. This characteristic will be explored further in the Results & Discussion section. We initially explored the possibility of fitting these bright point profiles to Gaussian functions in space and time. This was unsuccessful due to noise, the complexity of the bright points (particularly considering fragmentation), the high numbers of parameters needed to fully describe a moving Gaussian shape in 3 dimensions, and the small number of detected voxels for some candidate events. Furthermore, Vissers et al. (2015) suggest that a Gaussian model may not be appropriate for all TR phenomena.

Synthetic datacube
We apply the method to the synthetic datacube. The results are compared to the known distribution of the synthetic brightenings as a test of the method, and used to optimise various parameters. There are two ways to assess our method using synthetic data: a voxel-by-voxel set of statistics, and an event-by-event comparison. For the voxel-by-voxel statistics, we calculate the following: • True Positive (TP) -the number of detected voxels which correspond to synthetic brightening voxels • False Positive (FP) -the number of detected voxels which do not correspond to synthetic brightening voxels • False Negative (FN) -the number of non-detected voxels that correspond to synthetic brightening voxels • True Negative (TN) -the number of non-detected voxels that correspond to voxels without synthetic brightenings Note that we define a synthetic brightening voxel as voxels that have a brightness greater than 30% of their respective brightening's maximum amplitude, calculated from each brightening's Gaussian parameters. These criteria also define the Precision and Accuracy percentages, which are determined as: For an event-by-event comparison, a simple criterion we use is the distance of a detection's centroid to the closest true brightening, ∆ = ∆x 2 + ∆y 2 + ∆t 2 . This criterion allows a percentage scoring of the method with regards to accurate event detection, defined as where n good is the number of detections with the stringent criterion ∆ ≤ 10, and n total is the total number of brightenings in the synthetic datacube. Detections with ∆ > 10 are considered spurious. Table 1 shows various values of these criteria based on different values of the f low , f high , T low and T high parameters. For the synthetic data, we find an optimal band-pass at f low = 0.09, f high = 0.40 with threshold values of T low = 7 and T high = 9, whereby the detection percentage mentioned above reaches ∼ 81% and the TP detection reaches ∼ 50% while FPs remain < 10% and event-by-event accuracy remains high. The rest of this work is based on these parameters.
For 'true' event detections (∆ ≤ 10), figure 3a shows that there is a good correspondence between the detected maximum brightness and the true maximum brightness, with a mean absolute relative deviation (MARD) of 10.32% and a gradient of 1.13. Figure 3b compares the detected and true speed of the bright point, with a gradient of 1.04 and a MARD of 130.81%. Figure 3c shows the detected and true total brightness with a gradient of 0.32 and MARD 59.60%. This gradient of 0.32 shows that the total brightness is greatly underestimated. Since the maximum brightness of an event is accurately found, this underestimation is due to non-detection of brightening voxels, or an underestimation of a brightening's volume. This can be improved by decreasing T low , but at the expense of increasing false positives -our choice of parameters is therefore a compromise. Figure  3d shows a histogram of ∆ values, ∼ 99% of which are ≤ 3. These comparisons give approximate estimates of the uncertainty levels which we can assign to the brightnesses of detections in real data: the uncertainty of maximum brightness amplitude is small compared to general radiometric calibration errors, whilst the uncertainty in the total brightness is large, and comparable to the calibration errors.
All synthetic detections are dominated by N f rag = 1 values, with some detections with N f rag = 2. In an analysis of real data, a user can choose N f rag = 1 values only if they want to exclude overlapping or dubious detections. However, as stated in section 2, analysing the distribution of N f rag gives an idea of the complexity of these events.

A more complicated dataset
A second synthetic datacube is created containing ≈ 10 4 brightenings. The area, duration and maximum brightness of these brightenings follow continuous power-law distributions. The lower limits of these distributions are set at values below that which we expect to detect. The lower limit for the maximum brightness amplitude of the Gaussian events is 10 DN which is equal to the minimum 2σ noise level of the background. The lower limit for the spatial 1σ width is 0.4 pixels, and the lower limit for the temporal 1σ width is 0.25 time steps. The slopes of the power law distributions, as set by the standard parameter α (see (e.g. Clauset et al. 2007)), are α B = 2, α A = 3, and α t = 2.5, for brightness, area and duration respectively. These α values are set so that the largest values of each property are at expected values. That is, so the events do not become too large or long-lived. These power laws are shown as the black lines in figure 4. We assume that there is an approximate linear relationship between brightness, area and duration. For example, we expect the brightest events to possess, in general, a larger area and duration. We use a local randomisation process to impose these approximate relationship from the initial distributions, with the relationships shown in the plots of figure 5. The events have a uniform distribution of speeds ranging from zero to two pixels per timesteps, in random directions, and are placed in random positions throughout the datacube. Due to their speed and position, some events may appear or disappear at the datacube edges over time, and many events may overlap. In the case of overlaps, the maximum brightening at a given pixel is retained. This datacube is likely more representative of real-world IRIS event distributions and is a more realistic test for the detection method, whereby a large number of events are difficult or impossible to distinguish from noise, and many other events will not meet the filtering criteria i.e. some events will be too small or large, too faint, or too short-/long-lived. The same filtering process and parameters described in the previous sections are applied to this datacube. ∼ 23% of the 10,000 events are detected. For these detections, 88% are recorded as one coherent event (N f rag = 1), and 9% split into two regions at least once during their detected lifetimes (N f rag = 2).
The resulting detected power laws are shown in red in figure 4. Figures 4a shows a reasonable agreement between the slopes of the input/detected maximum brightness distributions only above ≈ 160DN. Below this limit, many events are detected, but at a small fraction of the true number. This fraction decreases rapidly with decreasing brightness. Above this limit, the number of detected events is slightly larger than the true number, showing an overestimation of brightness for some events. This is likely due to overlaps between brightenings, or an underestimation of the background brightness. A similar distribution holds for the total brightness. In analysing distributions of detections in real data, this decrease in detection rate for dimmer events must be properly modelled, for example following the approach of Clauset et al. (2007). Without accounting for the efficiency of the method to various brightness values, the power law slope α will always be greatly underestimated. This will be explored further in future work. Although detections may not directly correspond to input synthetic brightenings, the distribution of the detections still depends on the distribution of the synthetic inputs, even if the relationship becomes nonlinear. The nature of the detection process fundamentally changes the distribution of detections in a stochastic fashion and therefore a one-toone relationship between true events and those characterized by detection algorithms is not always possible.
Additionally, this study is limited to the 1400Å channel. It is possible and indeed likely that these brightening events could be emitting in several other wavelengths, which would suggest that energy output measurement are systematically underestimated by using only one channel. Indeed, some events may only exist at higher temperatures and energies, and this will have a direct impact on power-law gradients. This is similar to the arguments given by Aschwanden & Parnell (2002) and Aschwanden & Charbonneau (2002) in the context of nanoflares at much higher temperature ranges.
The distribution of detected durations shown in figure 4c follows the distribution of input durations well for events with σ time width of ≈1 or longer. Some event's durations are overestimated -this is likely due to overlaps between brightenings. No events with σ time widths of less than 0.6 are detected -this is due to the method discarding detections below this limit in order to reduce the number of false positives. The area distribution of figure 4d follows the input distribution well above a limit of ≈ 0.9 pixel σ area, with some event areas overestimated. The method does detect many small-area events below this limit, but at a fraction that decreases with decreasing area. While the detected distribution and input distribution match well down to a lower limit, this does not necessarily imply a one-to-one correspondence between a match and its true value. The slope of each distribution does not depend on the exact value of any detection (due to the scale-invariant nature of power-laws); only how they are distributed. Figure 6 shows a one-to-one comparison of input and detected events. The distance (or ∆, as defined previously) between all input events and a detection is calculated, and the input event with the smallest ∆ is paired to that detection. For comparison, we only consider the ∼ 11% of input events with ∆ ≤ 3. The distribution of ∆ for these selected pairs is shown in figure 6a. To statistically compare the detected and input properties, we fit the scattered points to a straight line using a leastabsolute-deviation (LAD) fitting procedure, and calculate the median relative deviation (MedRD, which will be positive/negative for systematic over/underestimation) and the median absolute relative deviation (MedRAD, which is the spread of deviations from the MedRD). The LAD fitted lines are shown in black in figures 6b-f. Note that the fitting is done in log space for 6c and d, and in linear space otherwise. The blue lines in 6b-f indicate the exact one-to-one relationship. Figure 6b compares the input and detected areas for the paired brightenings. The detected areas tend to underestimate the true areas with a MedRD of −15%, with an increasing underestimation with increasing area shown by the LAD fit gradient of 0.5. The scatter of points is high, with a MedRAD of 30%. The true log brightness shown in figure 6d shows an overwhelming underestimation of log brightness for the bulk of events with a MedRD value of −30%. The gradient of 1.1 shows that the larger/brighter events give a better estimate of total brightness, with a large scatter for dim/small events and an overall MedRAD of 22%.
The durations shown in figure 6e are similar to the areas -a tendency to underestimate (MedRD of −15%), with the underestimation increasing with increasing duration (LAD gradient of 0.5), and a large scatter (MedRAD of 25%). Figure 6f shows the performance of the method in tracking the centroid of events over time through comparing average speeds. The comparison is reasonable given the challenging task. The detected speeds increasingly underestimates the speed with increasing speed, with a gradient of 0.6, a MedRD of −16% and a MedRAD of 23%.
These comparisons will be central in interpreting the results, and power laws in particular, from real data in future studies, and highlight the difficulty of analysing power laws of small-scale brightenings in solar data. The one-to-one comparisons (only possible of course on synthetic data) show that only the detected maximum brightness can be considered a reliable measure for the brightest and most accurately detected events with a low uncertainty of ≈6%. With real data, one way of confirming whether a detection is true or not is to calculate ∆ across several wavelength channels. If an event appears across several channels and these facets are within an acceptable proximity to each other then the event is deemed to be true and the maximum brightness values recorded for those events are reliable. This will be implemented in a subsequent paper. The other values are mainly useful for showing trends or relationships for large statistical samples, with a systematic underestimation of area, duration, and total brightness, and a general statistical uncertainty of around 30%.

IRIS data
We analyse one data set from IRIS (De Pontieu et al. 2014) centred at approximately X = −126.05 ′′ , Y = 5.87 ′′ . The observations begun on Oct. 5th 2013 at 08:17 UT and ends the same day at 08:36. Figure 7 (left) shows the region of interest (ROI) as a QS region. The IRIS data consist of a 1400Å sit-and-stare series of slit-jaw images, with a consistent 9.46 s temporal cadence, 0.17 ′′ pixel −1 spatial scale and a 57.89 ′′ × 58.56 ′′ field of view (FOV). The IRIS data set is processed using standard procedures to account for temporal exposure, and empty margins are removed. Dark current subtraction as well as flat field, geometrical and orbital variation corrections have been applied as standard procedure for level 2 IRIS data. Isolated pixels with spuriously high values are flagged as missing data, and are ignored in the filtering and detection method. The central slit and its surrounding pixels are also treated as missing data such that extremely 'dark' pixels do not affect the filtering process. Additionally, drift due to solar rotation has been treated using a Fourier Local Correlation Tracking method (Fisher & Welsch 2008).
The results of the detection method with the IRIS datacube can be seen in figure 7, whereby the left plot represents the first image of the IRIS datacube. The right plot is the same IRIS image with all brightenings detected in the same time frame as the image on the left. Yellow contours represent brightenings with N f rag ≤ 2, green represent N f rag = 3 or N f rag = 4, and blue represents N f rag ≥ 5.  Figure 7. IRIS 1400Å slit-jaw images of (left) the ROI with (right) an over-plot of all detections from the same time frame. The left image has been processed such that isolated pixels or small regions of outlying intensity (such as the central slit) are treated as missing data. An animation of this figure is available. The video begins on Oct. 5th 2013 at 08:17 UT. The video ends the same day at 08:36.  Figure 10. Basic properties of the detections from the Oct. 5th 2013 IRIS data. The left, centre and right panels plot histograms of average speed, average area and total brightness, respectively. The dotted vertical lines represent the mean of each property; 9.03 kms −1 , 0.32 arcsec 2 and 6300 DNs −1 for speed, area and total brightness, respectively.
to N f rag = 2, as can be seen in figure 8. This yields a combined event density of 3.96 × 10 −4 events per second per arcsec 2 . The remainder consist of N f rag values between 3 and 12. This distribution of fragmented detections is quite different to the test datacube of section 3.2, and shows that the brightening detections in IRIS data tend to be more incoherent and fragmented. A more detailed analysis can show whether this is due to a true property of brightenings, or due to some other effect (e.g. high time variability of the background brightness, on which brightening events   Figure 11. Log-log scatter plots showing relationships between (a) area and duration, (b) duration and total brightness, (c) area and total brightness, and (d) volume vs maximum brightness. The green lines represent a robust linear fit to all points. The red triangle shows the mean value of both variables.
are superimposed, or overlapping of events in high-activity regions). For the preliminary results presented here, we have chosen to only include brightenings with N f rag = 1.
The choice of f low and f high has substantially reduced both small-scale, rapid noise and large-scale, slow changes, as can be seen in figure 9. The vertical dashed lines represent the coordinates of a real detection according to its maximum brightness values. It is clear from these plots, similar plots made for other detection events, as well as values of ∆ for synthetic data, that the method is effective at identifying and accurately determining the location of detections. Figure 10 presents some preliminary statistical results of these detections's properties: average speed (left), spatial area (center) and total brightness (right). Most detections have a speed between 3 and 15 kms −1 (with a mean value of 9.03 kms −1 ), cover an approximate average area between 0.2 and 0.4 arcsec 2 (with a mean value of 0.32 arcsec 2 ) and have a total brightness of 1600 − 31600 DN (with a mean value of 6300 DN). Large and/or bright events are rarer than small and/or dim events, and the distributions are all approximately symmetrical. As mentioned in section 3.2, the detection process is likely limited by observational resolution which would explain the distribution drop-off for smaller/dimmer/slower events. Figure 10 (right)'s profile, above the mean log brightness of ∼3.8, suggests that these brightenings follow a power-law distribution. Distributions such as these will be studied in more detail in future work. Figure 11 shows log-log scatter plots of area vs duration (a), total brightness vs duration (b), total brightness vs area (c) and volume vs maximum brightness (d). The green line shows the result of a robust linear fit to the logarithmic values, giving gradients of 0.14 (a), 1.60 (b), 1.32 (c) and 0.45 (d). The red triangle represents the mean value of both variables within each plot. The shallow gradient in figure 11a between area and duration suggests that there is no relationship between these properties. However, this result may be limited by the IRIS temporal cadence; the detection process detects the duration of events in integer number of frames, which can be seen by the sudden jumps in values along the y-axis. The plots of figures 11b-d show a clear proportional relationship between duration with total brightness, area with total brightness, and volume with maximum brightness, respectively. The former two relationships are unsurprising considering that a measurement of total brightness is dependent on the number of voxels. The latter relationship suggests that larger events tend to be intrinsically brighter than smaller events. Figure 12 shows the power of the method in analysing the properties of a brightening over its lifetime. Figure 12a tracks the motion of the brightening, which has an average speed of 0.06 pixel s −1 . Figure 12b shows the change in total brightness over time (measured in DN). Figure 12c shows the change in area over time. Figure 12d shows the change of maximum brightness over time. We find that this profile is typical of a large number of brightenings which have the property N f rag = 1, with a coherent motion and smooth increases/decreases in area and brightness during the event's lifetime. While increases in brightness and area intuitively coincide, figure 12d suggests that the intrinsic brightness also increases and decreases (independent of the size of the event).
A more detailed analysis of these relationships, applied to this and other data sets, will be made in a future study. We can draw comparisons between these and other studies of small-scale brightening phenomena, such as: Hα Ellerman Bombs (Ellerman 1917), which demonstrate typical widths, lengths and durations of 0.63 ′′ , 0.35 ′′ and 120s, respectively (Nelson et al. 2017b); UV IRIS bombs of approximately 2 ′′ × 2 ′′ area (Chen et al. 2019), and; Ca II chromospheric brightenings lasting at least 1.5 minutes over a 2 ′′ × 2 ′′ area (Kuckein, C. et al. 2017 Vissers et al. (2015)'s Flaring Arch Filaments rather than larger and longer-duration phenomena, which may explain the motion of some of these brightenings. We note that most of these previous studies focus on spectroscopic analyses which will lead to differences between their results and those of this study.

CONCLUSIONS AND FUTURE WORK
We present a general method to detect small-scale brightenings in solar time-series imagery. The method is applied to IRIS data sets in the UV, but can be applied to any time-series image data given appropriate adjustment of a few parameters. For a simple test dataset of IRIS-like synthetic slit-jaw image time series with randomly-distributed, randomly-moving small-scale Gaussian brightenings, the filtering and detection method yields an event-by-event detection rate of ∼ 81% and a voxel-byvoxel detection rate of ∼ 49% (with a voxel-by-voxel FP rate of < 10%). The detected maximum brightness values provide good accuracy to the true values for the brightest and most accurately detected events. The speeds are also reliable, albeit with a higher scatter. Care must be taken with measurements of total brightness -our tests show a large underestimation (0.3 on average) of total brightness due to the number of fainter voxels not being detected. The method is also applied to a second datacube with a high density of events possessing a power law distribution of brightness, area and duration. This test demonstrates the difficulty of extracting power law properties from real data due to instrument and method sensitivities to a limited range of brightness, duration and area. However, even though detected properties can have a large scatter from the true values, the shape of the detection histograms largely match the original distributions over a certain range. It also shows that the method gives an estimate of event maximum brightness with low uncertainty. Other properties (area, duration, speed, total brightness) have a large uncertainty, but the method gives approximate proportional relationships over large statistical samples.
2997 brightenings are detected in real IRIS data within a 54.40 ′′ × 55.23 ′′ area spanning 19 minutes. 1340 of these detected events either remain unfragmented or fragment at least once over their lifetime. These detections yield an approximate event density of 3.96×10 −4 arcsec −2 s −1 . Most detections have an average speed between 3 and 15 kms −1 (with a mean value of 9.03 kms −1 ), cover an approximate area between 0.2 and 0.4 arcsec 2 (with a mean value of 0.32 arcsec 2 ) and emit a total brightness of 1600 − 31600 DN (with a mean value of 6300 DN). Linear fitting of brightening properties shows that there are proportional relationships between the brightenings' area with total brightness, duration with total brightness, and volume with maximum brightness. However, there seems to be little proportionality between the brightenings' area with duration: this may be limited by IRIS's temporal cadence. Preliminary frame-by-frame plotting of a brightening event demonstrates coherent motion and a change in total brightness, area and intrinsic brightness over its lifetime -this seems typical of a large number of brightenings.
Having demonstrated the method's performance, we plan to apply it to a larger sample of quiet-sun IRIS data. This future study will use all the IRIS channels, as well as co-aligned data from other EUV imaging instruments, allowing a statistical study of brightening properties across a range of temperatures/heights. We also plan to test the method on ground-based photospheric data taken in visible wavelengths. The Interactive Data Language (IDL) software described in this paper will be released to the community in the near future through the SolarSoft library.