819
Views
5
CrossRef citations to date
0
Altmetric
Applications and Case Studies

Poisson-FOCuS: An Efficient Online Method for Detecting Count Bursts with Application to Gamma Ray Burst Detection

, ORCID Icon, ORCID Icon & ORCID Icon
Received 27 Jul 2022, Accepted 05 Jul 2023, Published online: 06 Sep 2023

Abstract

Gamma ray bursts are flashes of light from distant, new-born black holes. CubeSats that monitor high-energy photons across different energy bands are used to detect these bursts. There is a need for computationally efficient algorithms, able to run using the limited computational resource onboard a CubeSats, that can detect when gamma ray bursts occur. Current algorithms are based on monitoring photon counts across a grid of different sizes of time window. We propose a new method, which extends the recently proposed FOCuS approach for online change detection to Poisson data. Our method is mathematically equivalent to searching over all possible window sizes, but at half the computational cost of the current grid-based methods. We demonstrate the additional power of our approach using simulations and data drawn from the Fermi gamma ray burst monitor archive. Supplementary materials for this article are available online.

1 Introduction

This work is motivated by the challenge of designing an efficient algorithm for detecting gamma ray bursts (GRBs) for CubeSats, such as the HERMES scientific pathfinder mission (Fiore et al. Citation2020). GRBs are short-lived bursts of gamma ray light caused by the catastrophic accretion of matter into newly formed black holes. Long GRBs are associated with the formation of black holes in the collapse of massive, rapidly rotating stars, whereas short GRBs are associated with coalescence events in neutron star binary systems. These bursts were first detected by satellites in the late 1960s (Klebesadel, Strong, and Olson Citation1973). At the time of writing there is considerable interest in detecting gamma ray bursts due to their association with gravitational wave events (Luongo and Muccino Citation2021). For example, during August 17, 2017, the combined detection of the short gamma ray burst GRB170817 and the gravitational wave event GW170817 constrained the source’s location to a region of about 1100 deg2, roughly the size of the Ursa Major constellation in the night sky (Abbott et al. Citation2017). This limited the number of candidate host galaxies to a pool small enough to identify an optical counterpart. A large broadband observation campaign started soon after, leading to insights into many aspects of gravity and astrophysics (Miller Citation2017).

Instruments detecting GRBs must operate in space, as gamma light wavelengths are absorbed by nitrogen and oxygen in the upper layers of the atmosphere. Because of this, multiple CubeSats missions are being deployed to study them in the coming years (Bloser et al. Citation2022). CubeSats are compact and therefore relatively cheap to launch into space, but have limited computational power on-board. One of these missions is HERMES (High Energy Rapid Modular Ensemble of Satellites), a mission whose first six units will be launched in near-equatorial orbits during 2023, aiming to build an all-sky monitor for GRBs and other high-energy astronomical transients (Fiore et al. Citation2020). The HERMES main scientific goal is to monitor the whole sky for GRBs and locate their source directions.

Raw data from a satellite consists of a data stream of photons impacting a detector. The time of a photon impact is recorded in units of microseconds since satellite launch. New photon impacts are recorded on the order of approximately one every 500 microseconds (Campana et al. Citation2020). A GRB is indicated by a short period of time with an unusually high incidence of photons impacting the detector. Ideally the satellite would detect each GRB, and for each burst it detects it then transmits the associated data to earth.

There are a number of statistical challenges associated with detecting GRBs. First, bursts can come from close or far away sources with different intrinsic luminosities, and can therefore be either very bright and obvious to observe or very dim and hard to pick out from other background sources. They can also impact the detector over a variety of timescales. shows two GRBs, one short and intense lasting a fraction of a second, one longer and less intense lasting about 10 sec. These bursts were taken from the FERMI GBM Burst catalogue (Axelsson et al. Citation2019). Bursts ranging from a fraction of a second to a few minutes are possible (Kumar and Zhang Citation2015). Second, less than one GRB is recorded per 24 hr on average (Von Kienlin et al. Citation2020), which is relatively rare in comparison to the velocity of the signal. The background rate at which the satellite detects photons also varies over time. This is due to both rotations of the spacecraft and features of the near-Earth radiation environment at different points in orbit, which leads to irregularly cyclic patterns. This variation is on timescales much larger (many minutes to days) than those on which bursts occur (milliseconds to seconds), and thus is able to be estimated separately from the bursts. gives an example of a background signal.

Fig. 1 Plots of two recorded gamma ray bursts from the FERMI GBM Burst catalogue, with photon counts binned into 0.2 sec and 2 sec intervals.

Fig. 1 Plots of two recorded gamma ray bursts from the FERMI GBM Burst catalogue, with photon counts binned into 0.2 sec and 2 sec intervals.

Fig. 2 Example 4 hr of background data from one detector, grouped into 10 sec bins to show background rate fluctuations.

Fig. 2 Example 4 hr of background data from one detector, grouped into 10 sec bins to show background rate fluctuations.

Finally, there are also computational challenges. For example, there is limited computational hardware on board the satellite, and additional constraints may arise due to limitations of the electrical power system and lack of heat dissipation (Fenimore et al. Citation2003). There is also a substantial computational and energy cost to transmitting data to earth, so only promising data should be sent. This means we require any detection system to have a very low false positive rate.

At a fundamental level, algorithmic techniques for detecting GRBs have gone unchanged through different generations of space-born GRB monitor experiments (Feroci et al. Citation1997; Paciesas et al. Citation1999; Fenimore et al. Citation2003; Meegan et al. Citation2009). As they reach a detector, high-energy photons are counted over a fundamental time interval and in different energy bands. Count rates are then compared against a background estimate over a number of pre-defined timescales (Li and Ma Citation1983). To minimize the chance of missing a burst due to a mismatch between the event activity and the length of the tested timescales, multiple different timescales are simultaneously evaluated. Whenever the significance of the excess count is found to exceed a threshold value for that timescale, a trigger is issued.

gives a simplified overview of such a detection system. As data arrives we need to both detect whether a gamma ray burst is happening, and update our estimates of the background photon arrival rate. Because of the high computational cost of transmitting data to earth after a detection, if an algorithm detects a potential gamma ray burst there is an additional quality assurance step to determine whether it should be transmitted. This step often includes checking that a gamma ray burst has been detected at two or more detectors on the satellite. The detection algorithm needs to be run at a resolution at which all gamma ray bursts are detectable. By comparison background re-estimation is only required once every second, and the quality assurance algorithm is only needed every time a potential gamma ray burst is detected. Thus, the majority of the computational effort is for the detection algorithm—and how to construct a statistically efficient detection algorithm with low computation is the focus of this article.

Fig. 3 A schematic of the detection system, with the arrow thickness corresponding to the relative velocities of data flows. Most of the computing requirements of the trigger algorithm are within the detection loop, highlighted in green.

Fig. 3 A schematic of the detection system, with the arrow thickness corresponding to the relative velocities of data flows. Most of the computing requirements of the trigger algorithm are within the detection loop, highlighted in green.

As mentioned, current practice for detecting a GRB is to compare observed photon counts with expected counts across a given bin width (Paciesas et al. Citation2012). However, the choice of bin width affects the ease of discovery of different sizes of burst. shows an example. The short burst in is easily detectable with a bin width of 0.2 sec, but lost to smoothing at a 2 sec bin width. In contrast, the burst in has a signal too small relative to the noise to be detectable at a 0.2 sec bin width, with the largest observation on the plot being part of the noise rather than the gamma ray burst. Only when smoothed to a bin width of 2 sec does the burst become visually apparent. Therefore, the bin width is first chosen small enough to pick up short bursts, and geometrically spaced windows of size 1, 2, 4, 8,…times the bin width, up to a maximum window size, are run over the data in order to pick up longer bursts.

This article develops an improved approach to detecting GRBs. First we show that using the Page-CUSUM statistic (Page Citation1954, Citation1955), and its extension to count data (Lucas Citation1985) is uniformly better than using a window-based procedure. These schemes require specifying both the pre-change and post-change behavior of the data—in our case, this equates to specifying the background rate of photon arrivals and the rate during the gamma ray burst. While it is reasonable to assume, for our application, that good estimates of the background photon arrival rate are available, specifying the photon arrival rate for the gamma ray burst is difficult due to their heterogeneity in terms of intensity. For detecting changes in mean in Gaussian data, Romano et al. (Citation2023) show how one can implement the sequential scheme of Page (Citation1955) simultaneously for all possible post-change means, and call the resulting algorithm FOCuS. Our detection algorithm involves the nontrivial extension of this approach to the setting of detecting changes in the rate of events for count data. It is based on modeling the arrival of photons on the detector as a Poisson process, and we thus call our detection algorithm Poisson-FOCuS.

Poisson-FOCuS is equivalent to checking windows of any length, with a modified version equivalent to checking windows of any length up to a maximum size. This makes it advantageous for detecting bursts near the chosen statistical threshold whose length is not well described by a geometrically spaced window. In addition, the algorithm we develop has a computational cost lower than the geometric spacing approach, resulting in a uniform improvement on the methods already used for this application with no required tradeoff. These advantages mean that the Poisson-FOCuS algorithm is currently planned to be used as part of the trigger algorithm of the HERMES satellites.

The improvement Poisson-FOCuS achieves over existing windows based methods addresses the aspect of trigger algorithms that has been shown to be most important for increasing power of detecting GRBs. As the computational resource on-board a satellite have increased, trigger algorithms have grown to support an increasing number of criteria, and it has been seen that the most important aspect of any detection procedure is the timescale over which the data is analyzed (McLean et al. Citation2004). This is in contrast to other possible aspects such as testing data accumulated over multiple, fine-grained energy bands outside the standard 50–300 keV energy band. This did not result in more GRBs being detected by Fermi-GBM, and was eventually turned off to ease the computational burden on the on-board computer (Paciesas et al. Citation2012).

While early software, such as Compton-BATSE (Gehrels et al. Citation1993), operated only a few different trigger criteria, a total of 120 are available to the Fermi-GBM (Meegan et al. Citation2009) and more than 800 to the Swift-BAT (McLean et al. Citation2004) flight software. Whilst in many cases, this growth in algorithm complexity did not result in additional GRB detection, better coverage of different timescales for GRBs did (Paciesas et al. Citation2012). During the first four years of Fermi-GBM operations, 135 out of 953 GRBs were detected only over timescales not represented by BATSE algorithms, most of which were over timescales larger than the maximum value tested by BATSE (1.024 sec) (Von Kienlin et al. Citation2014).

The outline of the article is as follows. In Section 2 we define the mathematical setup of the problem. Section 3 introduces the functional pruning approach, leading to an algorithm and computational implementation specified in Section 4. In Section 5 we give an evaluation of our method on various simulated data, and real data taken from the FERMI catalogue. The article ends with a discussion.

2 Modeling Framework

The data we consider take the form of a time-series of arrival times of photons. We can model the generating process for these points as a Poisson process with background parameter λ(t), defining anomalies as periods of time which see an increase in the arrival rate over background level. By identifying anomalies, we hope to detect the GRBs that may cause them.

Changes in the background rate, λ(t), over time may be due to factors such as the spacecraft’s motion and the orbital radiation environment. They exist on a greater timescale (minutes to days) than the region of time over which a GRB could occur (seconds). We assume that a good estimate of the current background rate λ(t) is available. To ease exposition we will first assume this rate is known and constant, and denote it as λ, before generalizing to the nonconstant background rate in Section 4.1. We discuss accounting for error in estimating the background rate in Section 5.2. No autocorrelation is present in our data when this change in background rate has been accounted for, see the supplementary material.

The standard approach to analyzing our data is to choose a small time interval, w, which is smaller than the shortest GRB that we want to detect. Then the data can be summarized by the number of photon arrivals in time bins of length w. We denote the data by x1,x2,, with xi denoting the number of arrivals in the ith time window. We use the notation xt+1:t+h=(xt+1,,xt+h) to denote the vector of observations between the (t+1)th and (t+h)th time window, and x¯t+1:t+h=1hi=t+1t+hxi, the mean of these observations. Under our model, if there is no gamma ray burst then each xi is a realization of Xi , an independent Poisson random variable with parameter λ. If there is a gamma ray burst then the number of photon arrivals will be Poisson distributed with a rate larger than λ. We make the assumption that a gamma ray burst can be characterized by a width, h, and an intensity μ>1 such that if the gamma ray burst starts at time t + 1 then xt+1,,xt+h are realizations of independent Poisson random variables Xt+1,,Xt+h with mean μλ. See for a visualization of an anomaly simulated directly from this model.

Fig. 4 A simulated example anomaly with intensity multiplier μ = 3 and duration h = 20 against a background λ = 2.

Fig. 4 A simulated example anomaly with intensity multiplier μ = 3 and duration h = 20 against a background λ = 2.

Our algorithm is primarily interested in reducing the computational requirements of constant signal monitoring. Therefore, our model considers a gamma ray burst as a uniform increase in intensity over its length, which does not take into account the unknown shape of a gamma ray burst. If a possible burst is found, an additional round of shape-based sanity checking requiring more computational resources can easily be performed prior to transmission back to Earth.

2.1 Window-based Methods and Detectability

If we assumed we knew the width of the gamma ray burst, h, then detecting it would correspond to testing, for each start time t, between the following two hypotheses:

  • H0: Xt+1,,Xt+hPoisson(λ).

  • H1: Xt+1,,Xt+hPoisson(μλ), for some μ>1.

We can perform a likelihood ratio test for this hypothesis. Let denote the log-likelihood for the data xt+1:t+h under our Poisson model with rate μλ. Then the standard (log) likelihood ratio statistic is LR=2{maxμ>1l(xt+1:t+h;μ)l(xt+1:t+h;1)}.

This is 0 if x¯t+1:t+hλ, otherwise LR=2hλ{x¯t+1:t+hλlog(x¯t+1:t+hλ)(x¯t+1:t+hλ1)}.

The LR statistic is a function only of the expected count hλ and the fitted intensity μ̂t+1:t+h:=x¯t+1:t+h/λ of the interval [t+1,t+h]. Alternatively, it can be written as a function only of the expected count hλ and the actual count hx¯t+1:t+h, which forms the fundamental basis for our algorithm.

In our application, thresholds for gamma ray burst detection are often set based on k-sigma events: values that are as extreme as observing a Gaussian observation that is k standard deviations above its mean. For our one-sided test, the distribution of the likelihood ratio statistic under the null is approximately a mixture of a point mass at 0 and a χ12 distribution, each with probability 1/2 (Wilks Citation1938). We will call a k-sigma event one where the likelihood-ratio statistic exceeds a threshold of k2. The definition of a k-sigma event is based on a test for a single anomaly with a specific start and endpoint. In practice, detection of gamma ray bursts is achieved through performing multiple tests, allowing for different start and endpoints of the anomalies. Furthermore, different detection methods may perform more or fewer tests—which can make their statistical performance for the same k-sigma level different. We return to this in Section 5 where we present results that relate the k-sigma level for different methods to average run length, a common measure of Type-I error rate in sequential testing.

Gamma ray bursts with a combination of high intensity μ̂t+1:t+h:=x¯t+1:t+hλ and long length, as quantified by the expected count hλ, will have higher associated likelihood ratio statistics and thus be easier to detect. shows regions in this two-dimensional space that correspond to detectable GRBs at different k-sigma levels. shows what happens when we set a fixed threshold and for computational reasons only check intervals of certain lengths. We rely on the fact that a slightly brighter burst will also trigger detection on a longer or shorter interval than optimal. This is the type of approach that current window-based methods take (Paciesas et al. Citation2012). We can see that, even with a grid of window sizes, we lose detectability if the true width of the GRB does not match one of the window sizes.

Fig. 5 Detectability of GRBs at different k-sigma levels. Shaded regions show values of hλ and μ̂t+1:t+h where the likelihood ratio exceeds k-sigma event thresholds for k = 3 (blue region), k = 5 (orange region) and k = 7 (green region) for a test that uses the correct value of h.

Fig. 5 Detectability of GRBs at different k-sigma levels. Shaded regions show values of hλ and μ̂t+1:t+h where the likelihood ratio exceeds k-sigma event thresholds for k = 3 (blue region), k = 5 (orange region) and k = 7 (green region) for a test that uses the correct value of h.

Fig. 6 Detectability of GRBs by the window method, for one window (left) or a grid of three windows (right). The orange shaded area shows the values of hλ and μ̂t+1:t+h where the likelihood ratio exceeds a 5-sigma threshold, and the blue shaded area shows the detectability region form . Dashed lines show expected count hλ over the window.

Fig. 6 Detectability of GRBs by the window method, for one window (left) or a grid of three windows (right). The orange shaded area shows the values of hλ and μ̂t+1:t+h where the likelihood ratio exceeds a 5-sigma threshold, and the blue shaded area shows the detectability region form Figure 5. Dashed lines show expected count hλ over the window.

2.2 Page-CUSUM for Poisson Data

As a foundation for our detection algorithm, consider the CUSUM (cumulative sum) approach of Page (Citation1955) that was adapted for the Poisson setting by Lucas (Citation1985). These methods can search for gamma ray bursts of unknown width but known size μ, differing from a window method that searches for gamma ray bursts of known width h and unknown size. To run our methods online it is useful to characterize a possible anomaly by its start point τ. We have our hypotheses for the signal at time T:

  • H0: There have been no anomalies, that is, X1,,XTPoisson(λ).

  • H1: There has been one anomaly, beginning at some unknown time τ, with known intensity multiplier μ>1, that is, X1,,Xτ1Poisson(λ) and Xτ,,XTPoisson(μλ).

Our LR statistic for this test is 0 if x¯τ:Tλμ1log(μ) for all τ, otherwise LR=max1τT[2(Tτ+1)λ{x¯τ:Tλlog(μ)(μ1)}].

We work with a test statistic, ST , that is half the likelihood ratio statistic for this test, and compare to a k-sigma threshold of k2/2. ST can be rewritten in the following form: ST=[max1τTt=τT(xtlog(μ)λ(μ1))]+, where we use the notation [·]+ to denote the maximum of the term · and 0. As shown in Lucas (Citation1985), ST can be updated recursively as S0=0,ST+1=[ST+xT+1logμλ(μ1)]+.

It is helpful to compare the detectability of GRBs using ST with their detectability using a window method. To this end, we introduce the following propositions (see related results in Basseville and Nikiforov Citation1993; Romano et al. Citation2023).

Proposition 1.

For some choice of μ against a background rate of λ, let ST be significant at the k-sigma level. Then there exists some interval [τ,T] with associated likelihood ratio statistic that is significant at the k-sigma level.

Proposition 2.

For any k, λ, and h there exists a μ and corresponding test statistic, ST , that relates directly to a window test of length h, and background rate λ as follows: if, for any t, the data xt+1:t+h is significant at the k-sigma level then St+h will also be significant at the k-sigma level.

Together these results show that Page’s method is at least as powerful as the window method for detecting a GRB at a fixed k-sigma level. Rather than implementing the window method with a given window size, we can implement Page’s method with the appropriate μ value (as defined by Proposition 2) such that any GRB detected by the window method would be detected by Page’s method. However, Page’s method may detect additional GRBs and these would be detected by the window method with some window size (by Proposition 1). In practice, as shown in , Page’s method provides better coverage of the search space.

Fig. 7 Detectability of GRBs by Page-CUSUM for a single μ value (left) and a grid of three μ values (right). Orange shaded area shows the values of hλ and μ̂t+1:t+h where the likelihood ratio exceeds a 5-sigma threshold; the green shaded area shows the detectability region for the corresponding window test as defined by Proposition 2 (left-hand plot); and the blue shaded area shows the detectability region from .

Fig. 7 Detectability of GRBs by Page-CUSUM for a single μ value (left) and a grid of three μ values (right). Orange shaded area shows the values of hλ and μ̂t+1:t+h where the likelihood ratio exceeds a 5-sigma threshold; the green shaded area shows the detectability region for the corresponding window test as defined by Proposition 2 (left-hand plot); and the blue shaded area shows the detectability region from Figure 5.

Although the Page-CUSUM approach is more powerful than a window-based approach, to cover the space completely it still requires specifying a grid of values for the intensity of the gamma ray burst. If the actual intensity lies far from our grid values we will lose power at detecting the burst.

3 Functional Pruning

To look for an anomalous excess of count of any intensity and width without having to pick a parameter grid, we consider computing the Page-CUSUM statistic simultaneously for all μ[1,). This can be achieved by considering the test statistic as a function of μ, ST(μ). That is, for each T, ST(μ) is defined for μ[1,), and for a given μ is equal to the value of the Page-CUSUM statistic for that μ.

By definition, ST(μ) is a pointwise maximum of curves representing all possible anomaly start points τ: ST(μ):=[max1τTt=τT[xtlog(μ)λ(μ1)]]+.

We can view this as ST(μ)=[max1τTCτ(T)(μ)]+, where each curve, Cτ(T)(μ), corresponds to half the likelihood ratio statistic for a gamma ray burst of intensity μ starting at τ, Cτ(T)(μ):=t=τT[xtlog(μ)λ(μ1)].

Each curve is parameterized by two quantities, as Cτ(T)(μ):=aτ(T)log(μ)bτ(T)(μ1), where aτ(T)=t=τTxt is the actual observed count and bτ(T)=t=τTλ=λ(Tτ+1) is the expected count on the interval [τ,T]. As we move from time T to time T + 1 there is a simple recursion to update these coefficients: aτ(T+1)=aτ(T)+xT+1,bτ(T+1)=bτ(T)+λ. These are linear and do not depend on τ, so the differences between any two curves are preserved with time updates.

We call Cτ(T)(μ)a logarithmic curve. shows examples of these logarithmic curves. The maximum of Cτ(T) is located at μ=aτ(T)/bτ(T), representing the likelihood ratio for a post-change mean μλ=x¯τ:T. If and only if aτ(T)>bτ(T) then the logarithmic curve will be positive for some μ>1. In this case we will define the root of the curve to be the unique μ*>1 such that Cτ(T)(μ*)=0.

Fig. 8 Three example logarithmic curves. The statistic ST(μ) is defined as the maximum of all logarithmic curves and the 0 line.

Fig. 8 Three example logarithmic curves. The statistic ST(μ) is defined as the maximum of all logarithmic curves and the 0 line.

3.1 Adding and Pruning Curves

For any two curves Cτi(T) and Cτj(T) at a given present time T, we will say that Cτi(T) dominates Cτj(T) if [Cτi(T)(μ)]+[Cτj(T)(μ)]+, μ[1,).

This is equivalent to saying that there is no value of μ such that the interval [τj,T] provides better evidence for an anomaly with intensity μ than [τi,T]. As the difference between curves is unchanged as we observe more data, this in turn means that for any future point TFT, the interval [τj,TF] will not provide better evidence than [τi,TF]. Therefore, the curve associated with τj can be pruned, that is, removed from our computational characterization of ST(μ).

The following gives necessary and sufficient conditions for one curve to be dominated by another.

Proposition 3.

Let Cτi(T) and Cτj(T) be curves that are positive somewhere on μ[1,), where τi<τj and Cτi(τj1) is also positive somewhere on μ[1,). Then Cτi(T) dominates Cτj(T) if and only if aτj(T)/bτj(T)aτi(τj1)/bτi(τj1) or equivalently aτj(T)/bτj(T)aτi(T)/bτi(T). Additionally, it cannot be the case that Cτj(T) dominates Cτi(T).

A formal proof is given in the supplementary material, but we see why this result holds by looking at . There we see that Cτi(T) dominates Cτj(T) precisely when Cτi(T) has both a greater slope at μ = 1 (which occurs when Cτi(τj1) is positive) and a greater root than Cτj(T), where (as shown in the proof) the root of a curve Cτ(T) is an increasing function of aτ(T)/bτ(T).

4 Algorithm and Theoretical Evaluation

Using Proposition 3 we obtain the Poisson-FOCuS algorithm, described in Algorithm 1. This algorithm stores a list of curves in time order by storing their associated a and b parameters, as well as their times of creation τ, which for the constant λ case can be computed as T+1b/λ.

On receiving a new observation at time T, these parameters are updated. If the observed count exceeds the expected count predicted by the most recent curve we also add a new curve which corresponds to a GRB that starts at time T. Otherwise we check to see if we can prune the most recent curve. This pruning step uses Proposition 3, which shows that if any currently stored curve can be pruned, the most recently stored curve will be able to be pruned. (Our pruning check does not does not need to be repeated for additional curves, as on average less than one curve is pruned at each timestep.)

The final part of the algorithm is to find the maximum of each curve, and check if the maximum of these is greater than the threshold. If it is, then we have detected a GRB. The start of the detected GRB is given by the time that the curve with the largest maximum value was added.

Algorithm 1:

Poisson-FOCuS for constant λ

Result: Startpoint, endpoint and test-statistic value for anomaly detected at a k-sigma threshold.

1 set threshold k2/2;

2 initialize empty curve list;

3 while anomaly not yet found do

4 TT+1;

5 get actual count XT ;

6 get expected count λ;//update curves:

7 for curve in curve list

8 [Cτ1(T1),,Cτn(T1)] do aτi(T)aτi(T1)+XT; bτi(T)bτi(T1)+λ;

9 end//add or prune curve:

10 if XT/λ>max[aτn(T)/bτn(T),1] then

11 add CT(T):aT(T)=XT,bT(T)=λ,τ=T to curve list;

12 else if aτn(T)/bτn(T)<max[aτn1(T)/bτn1(T),1] then

13 remove Cτn(T) from curve list;

14 end

//calculate maximum M:

15 for curve Cτi(T) in curve list do

16 if max(Cτi(T))>M then

17 Mmax(Cτi(T));

18 τ*τi

19 end

20 end

21 if M>k2/2 then

22 anomaly found on interval [τ*,T] with 2M>k;

23 end

24 end

4.1 Dealing with Varying Background Rate

Algorithm 1 deals with the constant λ case. If λ=λ(t) is not constant, but an estimate λT of λ(T) is available at each timestep T, we can apply the same principle but with a change in the definition of bτ(T). We now have bτ(T):=t=τTλt, the total expected count over the interval [τ,T]. For the algorithm, this only impacts how the coefficients are updated, with the new updates being: aτ(T+1)aτ(T)+XT+1,bτ(T+1)bτ(T)+λT+1. If we work with a nonhomogeneous Poisson process in this way, it is impossible to recover τ from the coefficients aτ(T) and bτ(T), so Cτ(T) must be stored as the triplet (τ,aτ(T),bτ(T)).

The Poisson-FOCuS algorithm gives us an estimate of the start point of a GRB by reporting the interval [τ*,T] over which an anomaly is identified. In our application, if the additional sanity checking indicates a GRB is present, the whole signal starting some time before τ* to some time after T is then transmitted from the satellite to Earth. After this has occurred, Poisson-FOCuS can restart immediately provided that a good background rate estimate is available.

4.2 Minimum μ Value

For our application there is an upper limit on the length of a gamma ray burst, and it makes sense to ensure we do not detect gamma ray bursts that are longer than this. To do so, we set an appropriate μmin, and additionally prune curves which only contribute to ST(μ) on 1<μ<μmin, by removing, or not adding, curves Cτ(T) to the list if Cτ(T)(μmin)0, that is aτ(T)bτ(T)μmin1logμmin.

We can choose μmin according to our threshold and the maximum expected time period we are interested in searching for bursts over, using the proof of Proposition 2 about detectability for the window-based method, as follows: (hλ)max=k22[μminlog(μmin)(μmin1)].

For a 5-sigma threshold, assuming a background rate of one photon every 500μ s, a maximum length of 1 min for a GRB would correspond to μmin=1.015.

4.3 Computational Cost Comparisons

Using a window method, the computational cost per window consists of: adding xT and λT to the window; removing xTh and λth from the window; calculating the test statistic and comparing to the threshold. Using Poisson-FOCuS, our computational cost per curve consists of: adding xT to aτ(T); adding λT to bτ(T); calculating the maximum of the curve and comparing to the threshold. The computational cost per curve is therefore roughly equal to the computational cost per window. Thus, when evaluating the relative computational cost of Poisson-FOCuS versus a window method we need only calculate the expected number of curves kept by the algorithm at each timestep, and compare against the number of windows used. We now give mathematical bounds on this quantity, as follows:

Proposition 4.

The expected number of curves kept by Poisson-FOCuS without μmin at each timestep T is [log(T)2,log(T)+12].

Proposition 5.

The expected number of curves kept by Poisson-FOCuS using some μmin>1 at each timestep is bounded.

For geometrically spaced windows, over an infinite horizon the number of windows used at each timestep T is [log2(T),log2(T)+1], and if a hmax is implemented then this will be bounded after a certain point. gives a comparison of the number of windows and expected number of curves, showing that although the bound from Proposition 5 is difficult to calculate, it is substantially below the corresponding bound on the number of windows. Therefore, Poisson-FOCuS provides the statistical advantages of an exhaustive window search at under half the computational cost of a geometrically spaced one.

Fig. 9 Comparisons of the number of windows and expected number of curves (average over 1000 runs) kept by FOCuS running over a signal with base rate λ = 100 using a 5-sigma threshold, with no constraint on length of GRB (left) and with hmax=1024, corresponding to μmin=1.02 (right).

Fig. 9 Comparisons of the number of windows and expected number of curves (average over 1000 runs) kept by FOCuS running over a signal with base rate λ = 100 using a 5-sigma threshold, with no constraint on length of GRB (left) and with hmax=1024, corresponding to μmin=1.02 (right).

5 Empirical Evaluation

5.1 Simulations of GRBs and Average Run Length Comparison

We have previously worked under the assumption that a threshold level of 5 sigma will be used. In practice, the threshold is a tunable parameter and will be chosen to give a tradeoff between detection sensitivity and number of false detections.

We can make a direct comparison between the sigma threshold and average run length, a standard metric in anomaly detection literature. The average run length is the expected time until we have a false detection if we simulate data under the null hypothesis. This can be calculated for both FOCuS and logarithmically spaced windows, letting us make comparisons between the two by choosing a slightly different sigma threshold based on equal average run lengths. The average run length for each algorithm does not easily translate into an average run length for detecting GRBs, as it does not account for the sanity checking step of the detection algorithm, which can require a GRB to be detected by multiple detectors; nor do the results we present account for any error in estimating the background rate.

The results are shown in . The run length is given in terms of number of observations, so would need to be multiplied by the choice of fundamental bin width, w, to be converted to time. The main message from these results is that for the same average run length we would need to have a threshold that is 0.1σ higher for FOCuS than for the methods that uses logarithmically spaced windows.

Fig. 10 Comparison between FOCuS and logarithmic window method showing the average run length at different sigma levels.

Fig. 10 Comparison between FOCuS and logarithmic window method showing the average run length at different sigma levels.

We now compare Poisson-FOCuS with a window based method on synthetic data that has been simulated to mimic known GRBs, but allowing for different intensities of burst. To simulate the data for a chosen known GRB at a range of different brightnesses, the photon stream of the GRB was converted into a random variable via density estimation. One draw from this random variable would give a photon impact time, and n independent draws sorted into time order would give a stream of photon impact times that well approximate the shape of the burst. These were then overlaid on a background photon stream to form a signal that was binned into fundamental time widths of 50ms, which was fed into either Poisson-FOCuS or a geometrically spaced window search. The maximum sigma-level that was recorded when passing over the signal with each method is then plotted for various different brightnesses n. To stabilize any randomness introduced by the use of a random variable for GRB shape, this was repeated 10 times with different random seeds common to both methods and the average sigma-level is plotted.

The extent to which Poisson-FOCuS provides an improvement in detection power depends on the size and shape of the burst, and in particular whether the most promising interval in the burst lines up well with the geometrically spaced window grid. For example, the burst illustrated in does not line up with this grid, and shows how Poisson-FOCuS provides an improvement in detection power of approximately 0.5σ for this shape of burst at various different brightnesses, far higher than the approximately 0.1σ increase in threshold required to give a similar average run length (see ). However, the shorter burst in clearly has a most promising interval of size 1 for this binning choice, which is covered exactly by the logarithmic window grid. Therefore, Poisson-FOCuS provides no improvement over the window grid.

Fig. 11 Plots of runs of FOCuS and window-based method on simulated GRB copies of different brightnesses. Left-hand column shows three example GRBs each simulated to have 1000 photons. Right-hand column shows corresponding largest value of the test statistic as a function of number of photons in GRB. In plot (f) the lines for the two methods almost perfectly overlay one another.

Fig. 11 Plots of runs of FOCuS and window-based method on simulated GRB copies of different brightnesses. Left-hand column shows three example GRBs each simulated to have 1000 photons. Right-hand column shows corresponding largest value of the test statistic as a function of number of photons in GRB. In plot (f) the lines for the two methods almost perfectly overlay one another.

5.2 Bias from Estimating Background Rate

When using Poisson-FOCuS on a dataset requiring background rate estimation, particular care needs to be taken with the choice of the estimator for the background rate. Small, consistent under-estimation of the background rate will be recorded as an anomaly over a long timescale. The ability of any online algorithm to give immediate detections requires a background estimate for time T to be available using only data from tT, and this can cause challenges with avoiding underestimating the background rate in periods where it is increasing. Furthermore, the presence of a GRB within the data being used to estimate background rate could destabilize the estimation, so robust methods are needed.

To show this effect, shows a portion of the same data as at a higher resolution. This hour of data was chosen because it contains a smooth rise in background rate, which can give rise to anomalies when using a biased background rate estimation method.

Fig. 12 An hour’s portion of the same data from at higher resolution of 50ms (blue). In black is the data at 10 sec resolution identical to that from but rescaled by 0.005x to fit the graph. In orange is a centered 3 min moving-average background estimate (linewidth increased for visual clarity).

Fig. 12 An hour’s portion of the same data from Figure 2 at higher resolution of 50ms (blue). In black is the data at 10 sec resolution identical to that from Figure 2 but rescaled by 0.005x to fit the graph. In orange is a centered 3 min moving-average background estimate (linewidth increased for visual clarity).

shows the sigma thresholds recorded by Poisson-FOCuS running over this hour’s worth of higher-resolution data using a background estimate of 3 min centered moving-average window (), and a 3 min uncentered moving-average window such that background estimates at time T use only data from tT (). The uncentered method has a large peak in the recorded statistical threshold as it passes over the signal. This is caused by the upward change in background rate, which results in a consistent underestimation of the background rate at time T when using data from just the period prior to T. This bias in background estimation is then interpreted as a very small anomaly over a very long time period.

Fig. 13 Plots of a run of FOCuS over the data using various background estimation methods and parameters: (a) using a unbiased estimator of the background rate; (b) using a biased estimator; (c) using the same biased estimator but with μmin=1.1.

Fig. 13 Plots of a run of FOCuS over the data using various background estimation methods and parameters: (a) using a unbiased estimator of the background rate; (b) using a biased estimator; (c) using the same biased estimator but with μmin=1.1.

While the effect of a biased background estimation method can be somewhat countered by setting a value for μmin>1 as in , careful consideration should be given to de-biasing the background estimation method in order to avoid false detections when the background rate is rising (see e.g., Crupi et al. Citation2023). For example, we show in the next section that reliable detection of GRBs can be obtained using an exponential smoother to estimate the background rate.

5.3 Application to FERMI Data

In the context of HERMES, Poisson-FOCuS is currently being employed for two different purposes. First, a trigger algorithm built using Poisson-FOCuS is being developed for on-board, online GRB detection. To date, a dummy implementation has been developed and preliminary testings performed on the HERMES payload data handling unit computer. Second, Poisson-FOCuS is being employed in a software framework intended to serve as the foundation for the HERMES offline data analysis pipeline (Crupi et al. Citation2023). In this framework, background reference estimates are provided by a neural network as a function of the satellite’s current location and orientation.

Since no HERMES CubeSats have been launched yet, testing has taken place over Fermi gamma ray burst monitor (GBM) archival data, looking for events which may have evaded the on-board trigger algorithm. The data used for the analysis were drawn from the Fermi GBM daily data, Fermi GBM trigger catalogue, and Ferm GBM untriggered burst candidates catalogue, all of which are publically available at NASA’s High Energy Astrophysics Science Archive Research Center HEASARC (Citation2022a, Citation2022b), and Bhat (Citation2021).

The algorithm was run over eight days of data, from 00:00:00 2017/10/02 to 23:59:59 2017/10/09 UTC time. This particular time frame was selected because, according to the untriggered GBM Short GRB candidates catalog, it hosts two highly reliable short GRB candidates which defied the Fermi-GBM online trigger algorithm. During this week the Fermi GBM algorithm was triggered by 11 different events. Six of these were classified as GRBs, three as terrestrial gamma ray flashes, one as a local particle event and one as an uncertain event. The Poisson-FOCuS algorithm was run over data streams from 12 sodium iodide GBM detectors in the energy range of 50–300 kiloelectron volts, which is most relevant to GRB detection but excludes the bismuth germanate detectors and higher energy ranges designed to find terrestrial gamma ray flashes.

The data were binned at 100ms. Background count-rates were assessed by exponential smoothing of past observations, excluding the most recent 4 sec, and any curves corresponding to start points older than 4 sec were automatically removed from the curve lists. The returning condition used was the same used by Fermi-GBM: a trigger is issued whenever at least two detectors are simultaneously above threshold. After a trigger, the algorithm was kept idle for 5 min and then restarted.

At a 5-sigma threshold, Poisson-FOCuS was able to identify all the six GRBs which also triggered the Fermi-GBM algorithm, one of which is shown in . We also observed a trigger compatible with an event in the untriggered GBM Short GRB candidates catalog (Bhat Citation2021), which is shown in . An uncertain event not in either catalogue is shown in , which may indicate a GRB that had been missed by earlier searches.

Fig. 14 Three of the triggers found in the FERMI daily data. Left-hand column shows data from the two detectors that give a trigger, and the right-hand column shows the corresponding output from the Poisson-FOCuS algorithm.

Fig. 14 Three of the triggers found in the FERMI daily data. Left-hand column shows data from the two detectors that give a trigger, and the right-hand column shows the corresponding output from the Poisson-FOCuS algorithm.

6 Discussion

The main purpose of this article was to present a GRB detection algorithm that is mathematically equivalent to searching all possible window lengths, while requiring less computational power than the grid of windows approach. This algorithm is suitable for use on the HERMES satellites, where it has led to a reduction of required computations in a very computationally constrained setting, as well as a simplification of parameter choices by practitioners as values for window lengths in a grid no longer need to be specified.

There is increasing interest in detecting anomalies in other low-compute settings, for example Internet of Things sensors which must continuously monitor a signal (Dey et al. Citation2018). These may have a limited battery life or limited electricity generation from sensor-mounted solar panels. Therefore, the algorithm we have developed may be of use more widely.

Much of the mathematical work presented in this article is also applicable to the μ[0,1] case that searches for an anomalous lack of count in a signal. When adapting Poisson-FOCuS to this setting, it is important to make sure the algorithm functions well in situations where the counts are small, as these are precisely the locations of anomalies. Combining these two cases would give a general algorithm for detection of anomalies on μ[0,).

Supplemental material

UASA_A_2235059_Supplemental.zip

Download Zip (329.1 KB)

Supplementary Materials

Supplementary material contains proofs and empirical results showing the lack of auto-correlation in the GRB data.

Data Availability Statement

Code for Poisson-FOCuS and the analysis for this article is available at the GitHub repository https://github.com/kesward/FOCuS.

Disclosure Statement

No potential conflict of interest was reported by the authors.

Additional information

Funding

This work was supported by the EPSRC grants EP/N031938/1 and EP/R004935/1, and BT as part of the Next Generation Converged Digital Infrastructure (NG-CDI) Prosperity Partnership. We also acknowledge the support of the INAF RSN-5 mini-grant 1.05.12.04.05 “Development of novel algorithms for detecting high-energy transient events in astronomical time series”.

References

  • Abbott, B. P., Abbott, R., Abbott, T. D., Acernese, F., Ackley, K., Adams, C., et al. (2017), “Multi-Messenger Observations of a Binary Neutron Star Merger,” The Astrophysical Journal, 848, L12. DOI: 10.3847/2041-8213/aa91c9.
  • Axelsson, M., Bissaldi, E., Omodei, N., et al. (2019), “A Decade of Gamma-Ray Bursts Observed by Fermi-LAT: The Second GRB Catalog,” The Astrophysical Journal, 878, 52. DOI: 10.3847/1538-4357/ab1d4e.
  • Basseville, M., and Nikiforov, I. V. (1993), Detection of Abrupt Changes: Theory and Application, Englewood Cliffs: Prentice Hall.
  • Bhat, N. (2021), “Fermi Gamma-Ray Burst Monitor Untriggered GBM Short GRB Candidates,” available at https://gammaray.nsstc.nasa.gov/gbm/science/sgrb_search.html. Accessed: 2022-05-18.
  • Bloser, P. F., Murphy, D., Fiore, F., and Perkins, J. (2022), “Cubesats for Gamma-Ray Astronomy,” arXiv preprint arXiv:2212.11413.
  • Campana, R., Fuschino, F., Evangelista, Y., Dilillo, G., and Fiore, F. (2020), “The HERMES-TP/SP Background and Response Simulations,” in Space Telescopes and Instrumentation 2020: Ultraviolet to Gamma Ray (Vol. 11444), pp. 817–824, SPIE. DOI: 10.1117/12.2560365.
  • Crupi, R., Fiore, F., et al. (2023), “A Data Driven Framework for Searching Faint Gamma-Ray Bursts,” Unpublished.
  • Dey, N., Hassanien, A., Bhatt, C., Ashour, A., and Satapathy, S. (2018), Internet of Things and Big Data Analytics Toward Next-Generation Intelligence, Cham: Springer.
  • Fenimore, E., Palmer, D., Galassi, M., Tavenner, T., Barthelmy, S., Gehrels, N., Parsons, A., and Tueller, J. (2003), “The Trigger Algorithm for the Burst Alert Telescope on SWIFT,” AIP Conference Proceedings 662, 491–493.
  • Feroci, M., Frontera, F., Costa, E., Fiume, D. D., Amati, L., Bruca, L., et al. (1997), “In-Flight Performances of the BeppoSAX Gamma-Ray Burst Monitor,” in EUV, X-Ray, and Gamma-Ray Instrumentation for Astronomy VIII (Vol. 3114), pp. 186–197, SPIE.
  • Fiore, F., Burderi, L., Lavagna, M., Bertacin, R., Evangelista, Y., Campana, R., et al. (2020), “The HERMES-Technologic and Scientific Pathfinder,” in Space Telescopes and Instrumentation 2020: Ultraviolet to Gamma Ray (Vol. 11444), pp. 214–228.
  • Fiore, F., et al. (2020), “High Energy Rapid Modular Ensemble of Satellites Scientific Pathfinder,” available at https://www.hermes-sp.eu/. Accessed: 2021-08-03.
  • Gehrels, N., Fichtel, C. E., Fishman, G. J., Kurfess, J. D., and Schönfelder, V. (1993), “The Compton Gamma Ray Observatory,” Scientific American, 269, 68–77. DOI: 10.1038/scientificamerican1293-68.
  • HEASARC. (2022a), “Fermi GBM Daily Data,” available at https://heasarc.gsfc.nasa.gov/W3Browse/fermi/fermigdays.html. Accessed: 2022-06-07.
  • HEASARC. (2022b), “Fermi GBM Trigger Catalog,” available at https://heasarc.gsfc.nasa.gov/W3Browse/fermi/fermigtrig.html. Accessed: 2022-06-07.
  • Klebesadel, R. W., Strong, I. B., and Olson, R. A. (1973), “Observations of Gamma-Ray Bursts of Cosmic Origin,” The Astrophysical Journal, 182, L85–L88. DOI: 10.1086/181225.
  • Kumar, P., and Zhang, B. (2015), “The Physics of Gamma-Ray Bursts & Relativistic Jets,” Physics Reports, 561, 1–109. DOI: 10.1016/j.physrep.2014.09.008.
  • Li, T. P., and Ma, Y. Q. (1983), “Analysis Methods for Results in Gamma-Ray Astronomy,” The Astrophysical Journal, 272, 317–324. DOI: 10.1086/161295.
  • Lucas, J. M. (1985), “Counted Data CUSUMs,” Technometrics, 27, 129–144. DOI: 10.1080/00401706.1985.10488030.
  • Luongo, O., and Muccino, M. (2021), “A Roadmap to Gamma-Ray Bursts: New Developments and Applications to Cosmology,” Galaxies, 9, 77. DOI: 10.3390/galaxies9040077.
  • McLean, K., Fenimore, E. E., Palmer, D., Barthelmy, S., Gehrels, N., Krimm, H., Markwardt, C., and Parsons, A. (2004), “Setting the Triggering Threshold on Swift,” AIP Conference Proceedings, 727, 667–670.
  • Meegan, C., Lichti, G., Bhat, P. N., Bissaldi, E., Briggs, M. S., Connaughton, V. (2009), “The Fermi Gamma-Ray Burst Monitor,” The Astrophysical Journal, 702, 791. DOI: 10.1088/0004-637X/702/1/791.
  • Miller, M. C. (2017), “A Golden Binary,” Nature, 551, 36–37. DOI: 10.1038/nature24153.
  • Paciesas, W. S., Meegan, C. A., Pendleton, G. N., Briggs, M. S., Kouveliotou, C., and Koshut, T. M. et al. (1999), “The Fourth BATSE Gamma-Ray Burst Catalog (revised),” The Astrophysical Journal Supplement Series, 122, 465. DOI: 10.1086/313224.
  • Paciesas, W. S., Meegan, C. A., von Kienlin, A., Bhat, P. N., Bissaldi, E., Briggs, M. S., Michael Burgess, J., et al. (2012), “The Fermi GBM Gamma-Ray Burst Catalog: The First Two Years,” The Astrophysical Journal Supplement Series, 199, 18. DOI: 10.1088/0067-0049/199/1/18.
  • Page, E. S. (1954), “Continuous Inspection Schemes,” Biometrika, 41, 100–115.
  • Page, E. S. (1955), “A Test for a Change in a Parameter Occurring at an Unknown Point,” Biometrika, 42, 523–527.
  • Romano, G., Eckley, I. A., Fearnhead, P., and Rigaill, G. (2023), “Fast Online Changepoint Detection via Functional Pruning CUSUM Statistics,” Journal of Machine Learning Research, 24, 1–36. http://jmlr.org/papers/v24/21-1230.html
  • Von Kienlin, A., Meegan, C. A., Paciesas, W. S., Bhat, P., Bissaldi, E., Briggs, M. S., Burgess, J. M., Byrne, D. et al. (2014), “The Second Fermi GBM Gamma-Ray Burst Catalog: The First Four Years,” The Astrophysical Journal Supplement Series, 211, 13. DOI: 10.1088/0067-0049/211/1/13.
  • Von Kienlin, A., Meegan, C. A., Paciesas, W. S., Bhat, P., Bissaldi, E., Briggs, M. S., et al. (2020), “The Fourth Fermi-GBM Gamma-Ray Burst Catalog: A Decade of Data,” The Astrophysical Journal, 893, 46. DOI: 10.3847/1538-4357/ab7a18.
  • Wilks, S. S. (1938), “The Large-Sample Distribution of the Likelihood Ratio for Testing Composite Hypotheses,” The Annals of Mathematical Statistics, 9, 60–62. DOI: 10.1214/aoms/1177732360.