Sample size determination for open-ended questions or qualitative interviews relies primarily on custom and finding the point where little new information is obtained (thematic saturation). Here, we propose and test a refined definition of saturation as obtaining the most salient items in a set of qualitative interviews (where items can be material things or concepts, depending on the topic of study) rather than attempting to obtain all the items. Salient items have higher prevalence and are more culturally important. To do this, we explore saturation, salience, sample size, and domain size in 28 sets of interviews in which respondents were asked to list all the things they could think of in one of 18 topical domains. The domainslike kinds of fruits (highly bounded) and things that mothers do (unbounded)varied greatly in size. The datasets comprise 2099 interviews each (1,147 total interviews). When saturation was defined as the point where less than one new item per person would be expected, the median sample size for reaching saturation was 75 (range = 15194). Thematic saturation was, as expected, related to domain size. It was also related to the amount of information contributed by each respondent but, unexpectedly, was reached more quickly when respondents contributed less information. In contrast, a greater amount of information per person increased the retrieval of salient items. Even small samples (n = 10) produced 95% of the most salient ideas with exhaustive listing, but only 53% of those items were captured with limited responses per person (three). For most domains, item salience appeared to be a more useful concept for thinking about sample size adequacy than finding the point of thematic saturation. Thus, we advance the concept of saturation in salience and emphasize probing to increase the amount of information collected per respondent to increase sample efficiency.
CHENGXINRUN are exported all over the world and different industries with quality first. Our belief is to provide our customers with more and better high value-added products. Let's create a better future together.
As modeling estimates to date have been based on only one or two real-world examples, it is clear that more empirical examples are needed. Here, we use 28 real-world examples to estimate the impact of sample size, domain size, and amount of information per respondent on saturation and on the total number of items obtained. Using the proportion of people in a sample that mentioned an item as a measure of salience, we find that even small samples may adequately capture the most salient items.
Similarly, Lowe et al. [ 25 ] estimated saturation and domain size in two examples and in simulated datasets, testing the effect of various parameters. Lowe et al. found that responses were not independent across respondents and that saturation may never be reached. In this context, non-independence refers to the fact that some responses are much more likely than others to be repeated across people. Instead of complete saturation, they suggested using a goal such as obtaining a percentage of the total domain that one would like to capture (e.g., 90%) and the average prevalence of items one would like to observe to estimate the appropriate sample size. For example, to obtain 90% of items with an average prevalence of 0.20, a sample size of 36 would be required. Van Rijnsoever [ 26 ] used simulated datasets to study the accumulation of themes across sample size increments and assessed the effect of different sampling strategies, item prevalence, and domain size on saturation. Van Rijnsoevers results indicated that the point of saturation was dependent on the prevalence of the items.
Saturation and sample size have also begun to be examined with multivariate models and simulations. Tran et al. [ 24 ] estimated thematic saturation and the total number of themes from open-ended questions in a large survey and then simulated data to test predictions about sample size and saturation. They assumed that items were independent and found that sample sizes greater than 50 would add less than one new theme per additional person interviewed.
Galvin[ 21 ] and Fugard and Potts[ 22 ] framed the sample size problem for qualitative data in terms of the likelihood that a specific idea or theme will or will not appear in a set of interviews, given the prevalence of those ideas in the population. They used traditional statistical theory to show that small samples retrieve only the most prevalent themes and that larger samples are more sensitive and can retrieve less prevalent themes as well. This framework can be applied to the expectation of observing or not observing almost anything. Here it would apply to the likelihood of observing a theme in a set of narrative responses, but it applies equally well for situations such as behavioral observations, where specific behaviors are being observed and sampled[ 23 ]. For example, to obtain ideas or themes that would be reported by about one out of five people (0.20 prevalence) or a behavior with the same prevalence, there is a 95% likelihood of seeing those themes or behaviors at least once in 14 interviewsif those themes or behaviors are independent.
Increasingly, researchers are applying systematic analysis and sampling theory to untangle the problems of saturation and sample size in the enormous variety of studies that rely on qualitative dataincluding life-histories, discourse analysis, ethnographic decision modeling, focus groups, grounded theory, and more. For example, Guest et al.[ 17 ] and others[ 18 19 ] found that about 1216 interviews were adequate to achieve thematic saturation. Similarly, Hagaman and Wutich [ 20 ] found that they could reliably retrieve the three most salient themes from each of the four sites in the first 16 interviews.
In this article, we estimate the point of complete thematic saturation and the associated sample size and domain size for 28 sets of interviews in which respondents were asked to list all the things they could think of in one of 18 topical domains. The domainslike kinds of fruits (highly bounded) and things that mothers do (unbounded)varied greatly in size. We also examine the impact of the amount of information produced per respondent on saturation and on the number of unique items obtained by comparing results generated by asking respondents to name all the relevant things they can with results obtained from a limited number of responses per question, as with standard open-ended questioning. Finally, we introduce an additional type of saturation based on the relative salience of items and themessaturation in salienceand we explore whether the most salient items are captured at minimal sample sizes. A key conclusion is that saturation may be more meaningfully and more productively conceived of as the point where the most salient ideas have been obtained.
Since the goal of research based on qualitative data is not necessarily to collect all or most ideas and themes but to collect the most important ideas and themes, salience may provide a better guide to sample size adequacy than saturation. Salience (often called cultural or cognitive salience) can be measured by the frequency of item occurrence (prevalence) or the order of mention [ 7 , 8 ]. These two indicators tend to be correlated [ 9 ]. In a set of lists of birds, for example, robins are reported more frequently and appear earlier in responses than are penguins. Salient terms are also more prevalent in everyday language [ 10 12 ]. Item salience also may be estimated by combining an items frequency across lists with its rank/position on individual lists [ 13 16 ].
Open-ended questions are used alone or in combination with other interviewing techniques to explore topics in depth, to understand processes, and to identify potential causes of observed correlations. Open-ended questions may produce lists, short answers, or lengthy narratives, but in all cases, an enduring question is: How many interviews are needed to be sure that the range of salient items (in the case of lists) and themes (in the case of narratives) are covered. Guidelines for collecting lists, short answers, and narratives often recommend continuing interviews until saturation is reached. The concept of theoretical saturationthe point where the main ideas and variations relevant to the formulation of a theory have been identifiedwas first articulated by Glaser and Strauss [ 1 , 2 ] in the context of how to develop grounded theory. Most of the literature on analyzing qualitative data, however, deals with observable thematic saturationthe point during a series of interviews where few or no new ideas, themes, or codes appear [ 3 6 ].
A log-linear solution generalizes this logic from a 2 x 2 table to a 2 K table [ 34 ]. the capturerecapture solution estimates total population size for hidden populations using the pattern of recapture (matching) between pairs of samples (respondents) to estimate the population size. An implementation in R with GLM uses a log-linear form to estimate population size based on recapture rates (Rcapture [ 35 , 36 ]). In this application, it is assumed that the population does not change between interviews (closed population) and models are fit with: (1) no variation across people or items (M 0 ); (2) variation only across respondents (M t ); (3) variation only across items (M h ); and (4) variation due to an interaction between people and items (M ht ). For each model, estimates were fit with binomial, Chaos lower bound estimate, Poisson, Darroch log normal, and gamma distributions [ 35 ]. Variation among items (heterogeneity) is a test for a difference in the probabilities of item occurrence and, in this case, is equivalent to a test for a difference in item salience among the items. Due to the large number of combinations needed to estimate these models, Rcapture software estimates are provided for all four models only up to a sample of size 10. For larger sample sizes (all examples in this study had sample sizes of 20 or larger), only model 1 with no effects for people or items (the binomial model) and model 3 with item effects (item salience differences) were tested. Therefore, models were fit at size 10, to test all four models and then at the total available sample size.
Because the GLM models above used only incremental sample size to predict the total number of unique items (domain size) and ignored variation in the number of items provided by each person and variation in item salience, an additional analysis was used to estimate domain size while accounting for subject and item heterogeneity. For that analysis, domain size was estimated with a capture-recapture estimation technique used for estimating the size of hidden populations. Domain size was estimated from the total number of items on individual lists and the number of matching items between pairs of lists with a log-linear analysis. For example, population size can be estimated from the responses of two people as the product of their number of responses divided by the number of matching items (assumed to be due to chance). If Person#1 named 15 illness terms and Person#2 named 31 terms and they matched on five illnesses, there would be 41 unique illness terms and the estimated total number of illness terms based on these two people would be (15 x 31) /5 = 93.
To help develop an interviewing stopping rule, a simple model was used to predict the unique number of items contributed by each additional respondent. Generalized linear models (GLM, log-linear models for count data) were used to predict the unique number of items added by each respondent (incrementing sample size), because number of unique items added by each respondent (count data) is approximately Poisson distributed. For each example, models were fit with ordinary least squares linear regression, Poisson, and negative binomial probability distributions. Respondents were assumed to be in random order, in the order in which they occurred in each dataset, although in some cases they were in the order they were interviewed. Goodness-of-fit was compared across the three models with minimized deviants (the Akaike Information Criterion, AIC) to find the best-fitting model [ 33 ]. Using the best-fitting model for each example, the point of saturation was estimated as the point where the expected number of new items was one or less. Sample size and domain size were estimated at the point of saturation, and total domain size was estimated for an infinite sample size from the model for each example as the limit of a geometric series (assuming a negative slope).
For each example, statistical models describe the pattern of obtaining new or unique items with incremental increases in sample size. Individual lists were first analyzed with Flame [ 31 , 32 ] to provide the list of unique items for each example and the Smith [ 14 ] and Sutrop [ 15 ] item salience scores. Duplicate items due to spelling, case errors, spacing, or variations were combined.
Some interviews were face to face, some were written responses, and some were administered on-line. Investigators varied in their use of prompts, using nonspecific (What other are there?), semantic (repeating prior responses and then asking for others), and/or alphabetic prompts (going through the alphabet and asking for others). Brewer [ 29 ] and Gravlee et al. [ 30 ] specifically examined the effect of prompting on response productivity, although the Brewer et al. examples in these analyses contain results before extensive prompting and the Gravlee et al. examples contain results after prompting. The 28 examples, their topic, source, sample size, the question used in the original data collection, and the three most frequently mentioned items appear in . All data were collected and analyzed without personal identifying information.
Data were obtained by contacting researchers who published analyses of free lists. Examples with 20 or more interviews were selected so that saturation could be examined incrementally through a range of sample sizes. Thirteen published examples were obtained on: illness terms [ 27 ] (in English and in Spanish); birds, flowers, and fabrics [ 28 ]; recreational/street drugs and fruits [ 29 ]; things mothers do (online, face-to-face, and written administration) and racial and ethnic groups [ 30 ] (online, face-to-face, and written administration). Fifteen unpublished classroom educational examples were obtained on: soda pops (Weller, n.d.); holidays (two replications), things that might appear in a living room, characteristics of a good leader (two replications), a good team (two replications), and a good team player (Johnson, n.d.); and bad words, industries (two replications), cultural industries (two replications), and scary things (Borgatti, n.d.). (Original data appear online in S1 Appendix The Original Data for the 28 Examples.)
The datasets comprise 2099 interviews each (1,147 total interviews). Each example elicits multiple responses from each individual in response to an open-ended question (Name all the you can think of) or a question with probes (What other are there?).
Descriptive information for the examples appears in . The first four columns list the name of the example, the sample size in the original study, the mean list length (with the range of the list length across respondents), and the total number of unique items obtained. For the Holiday1 example, interviews requested names of holidays (Write down all the holidays you can think of), there were 24 respondents, the average number of holidays listed per person (list length) was 13 (ranging from five to 29), and 62 unique holidays were obtained.
The free-list counts showed a characteristic descending curve where an initial person listed new themes and each additional person repeated some themes already reported and added new items, but fewer and fewer new items were added with incremental increases in sample size. All examples were fit using the GLM log-link and identity-link with normal, Poisson, and negative binomial distributions. The negative binomial model resulted in a better fit than the Poisson (or identity-link models) for most full-listing examples, providing the best fit to the downward sloping curve with a long tail. Of the 28 examples, only three were not best fit by negative binomial log-link models: the best-fitting model for two examples was the Poisson log-link model (GoodTeam1 and GoodTeam2Player) and one was best fit by the negative binomial identity-link model (CultInd1).
Sample size was a significant predictor of the number of new items for 21 of the 28 examples. Seven examples did not result in a statistically significant fit (Illnesses-US, Holiday2, Industries1, Industries2, GoodTLeader, GoodTeam2Player, and GoodTeam3). The best-fitting model was used to predict the point of saturation and domain size for all 28 examples (S2 Appendix GLM Statistical Model Results for the 28 Examples).
Using the best-fitting GLM models we estimated the predicted sample size for reaching saturation. Saturation was defined as the point where less than one new item would be expected for each additional person interviewed. Using the models to solve for the sample size (X) when only one item was obtained per person (Y = 1) and rounding up to the nearest integer, provided the point of saturation (Y1.0). , column five, reports the sample size where saturation was reached (NSAT). For Holiday1, one or fewer new items were obtained per person when X = 16.98. Rounding up to the next integer provides the saturation point (NSAT = 17). For the Fruit domain, saturation occurred at a sample size of 15.
Saturation was reached at sample sizes of 15194, with a median sample size of 75. Only five examples (Holiday1, Fruits, Birds, Flowers, and Drugs) reached saturation within the original study sample size and most examples did not reach saturation even after four or five dozen interviews. A more liberal definition of saturation, defined as the point where less than two new items would be expected for each additional person (solving for Y2), resulted in a median sample size for reaching saturation of 50 (range 10146).
Some domains were well bounded and were elicited with small sample sizes. Some were not. In fact, most of the distributions exhibited a very long tailwhere many items were mentioned by only one or two people. shows the predicted curves for all examples for sample sizes of 1 to 50. Saturation is the point where the descending curve crosses Y = 1 (or Y = 2). Although the expected number of unique ideas or themes obtained for successive respondents tends to decrease as the sample size increases, this occurs rapidly in some domains and slowly or not at all in other domains. Fruits, Holiday1, and Illness-G are domains with the three bottom-most curves and the steepest descent, indicating that saturation was reached rapidly and with small sample sizes. The three top-most curves are the Moms-F2F, Industries1, and Industries2 domains, which reached saturation at very large sample sizes or essentially did not reach saturation.
Open in a separate windowBecause saturation appeared to be related to domain size and some investigators state that a percentage of the domain might be a better standard [25], domain size was also estimated. First, total domain size was estimated with the GLM models obtained above. Domain size was estimated at the point of saturation by cumulatively summing the number of items obtained for sample sizes n = 1, n = 2, n = 3, to NSAT. For the Holiday1 sample, summing the number of predicted unique items for sample sizes n = 1 to n = 17 should yield 51 items ( , Domain Size at Saturation, DSAT). Thus, the model predicted that approximately 51 holidays would be obtained by the time saturation was reached.
The total domain size was estimated using a geometric series, summing the estimated number of unique items obtained cumulatively across people in an infinitely large sample. For the Holiday1 domain, the total domain size was estimated as 56 (see , Total Domain Size DTOT). So for the Holiday1 domain, although the total domain size was estimated to be 57, the model predicted that saturation occurred when the sample size reached 17, and at that point 51 holidays should be retrieved. Model predictions were close to the empirical data, as 62 holidays were obtained with a sample of 24.
Larger sample sizes were needed to reach saturation in larger domains; the largest domains were MomsF2F, Industries1, and Industries2 each estimated to have about 1,000 items and more than 100 interviews needed to approach saturation. Saturation (Y1) tended to occur at about 90% of the total domain size. For Fruits, the domain size at saturation was 51 and the total domain size was estimated at 53 (51/53 = 96%) and for MomsF2F, domain size at saturation was 904 and total domain size was 951 (95%).
Second, total domain size was estimated using a capture-recapture log-linear model with a parameter for item heterogeneity [35,36]. A descending, concave curve is diagnostic of item heterogeneity and was present in almost all of the examples. The estimated population sizes using R-Capture appear in the last column of . When the gamma distribution provided the best fit to the response data, the domain size increased by an order of magnitude as did the standard error on that estimate. When responses fit a gamma distribution, the domain may be extremely large and may not readily reach saturation.
Inclusion of the pattern of matching items across people with a parameter for item heterogeneity (overlap in items between people due to salience) resulted in larger population size estimates than those above without heterogeneity. Estimation from the first two respondents was not helpful and provided estimates much lower than those from any of the other methods. The simple model without subject or item effects (the binomial model) did not fit any of the examples. Estimation from the first 10 respondents in each example suggested that more variation was due to item heterogeneity than to item and subject heterogeneity, so we report only the estimated domain size with the complete samples accounting for item heterogeneity in salience.
Overall, the capturerecapture estimates incorporating the effect of salience were larger than the GLM results above without a parameter for salience. For Fruits, the total domain size was estimated as 45 from the first two people; as 88 (gamma distribution estimate) from the first 10 people with item heterogeneity and as 67 (Chao lower bound estimate) with item and subject heterogeneity; and using the total sample (n = 33) the binomial model (without any heterogeneity parameters) estimated the domain size as 62 (but did not fit the data) and with item heterogeneity the domain size was estimated as 73 (the best-fitting model used the Chao lower bound estimate). Thus, the total domain size for Fruits estimated with a simple GLM model was 53 and with a capturerecapture model (including item heterogeneity) was 73 ( , last column). Similarly, the domain size for Holiday1 was estimated at 57 with the simple GLM model and 100 with capture-recapture model. Domain size estimates suggest that even the simplest domains can be large and that inclusion of item heterogeneity increases domain size estimates.
The original examples used an exhaustive listing of responses to obtain about a half dozen (GoodLeader and GoodTeam2Player) to almost three dozen responses per person (Industries1 and Industries2). A question is whether saturation and the number of unique ideas obtained might be affected by the number of responses per person. Since open-ended questions may obtain only a few responses, we limited the responses to a maximum of three per person, truncating lists to see the effect on the number of items obtained at different sample sizes and the point of saturation.
When more information (a greater number of responses) was collected per person, more unique items were obtained even at smaller sample sizes ( ). The amount of information retrieved per sample can be conceived of in terms of bits of information per sample and is roughly the average number of responses per person times the sample size so that, with all other things being equal, larger sample sizes with less probing should approach the same amount of information obtained with smaller samples and more probing. So, for a given sample size, a study with six responses per person should obtain twice as much information as a study with three responses per person. In the GoodLeader, GoodTeam1, and GoodTeam2Player examples, the average list length was approximately six and when the sample size was 10 (6 x 10 = 60 bits of information), approximately twice as many items were obtained as when lists were truncated to three responses (3 x 10 = 30 bits of information).
Increasing the sample size proportionately increases the amount of information, but not always. For Scary Things, 5.6 bits more information were collected per person with full listing (16.9 average list length) than with three or fewer responses per person (3.0 list length); and the number of items obtained in a sample size of 10 with full listing (102) was roughly 5.6 times greater than that obtained with three responses per person (18 items). However, at a sample size of 20 the number of unique items with free lists was only 4.5 times larger (153) than the number obtained with three responses per person (34). Across examples, interviews that obtained more information per person were more productive and obtained more unique items overall even with smaller sample sizes than did interviews with only three responses per person.
Using the same definition of saturation (the point where less than one new item would be expected for each additional person interviewed), less information per person resulted in reaching saturation at much smaller sample sizes. shows the predicted curves for all examples when the number of responses per person is three (or fewer). The Holiday examples reached saturation (fewer than one new item per person) with a sample size of 17 (Holiday1) with 13.0 average responses per person and 87 (Holiday2) with 17.8 average responses ( ), but reached saturation with a sample size of only 9 (Holiday 1 and Holiday2) when there were a maximum of three responses per person ( , last column). With three or fewer responses per person, the median sample size for reaching saturation was 16 (range: 4134). Thus, fewer responses per person resulted in reaching saturation at smaller sample sizes and resulted in fewer domain items.
Open in a separate windowSaturation did not seem to be a useful guide for determining a sample size stopping point, because it was sensitive both to domain size and the number of responses per person. Since a main goal of open-ended interviews is to obtain the most important ideas and themes, it seemed reasonable to consider item salience as an alternative guide to assist with determining sample size adequacy. Here, the question would be: Whether or not complete saturation is achieved, are the most salient ideas and themes captured in small samples?
A simple and direct measure of item salience is the proportion of people in a sample that mentioned an item [37]. However, we examined the correlation between the sample proportions and two salience indices that combine the proportion of people mentioning an item with the items list position [1315]. Because the item frequency distributions have long tailsthere are many items mentioned by only one or two peoplewe focused on only those items mentioned by two or more people (24204 items) and used the full lists provided by each respondent. The average Spearman correlation between the Smith and Sutrop indices in the 28 examples was 0.95 (average Pearson correlation 0.96, 95%CI: 0.92, 0.98), between the Smith index and the sample proportions was 0.89 (average Pearson 0.96, 95%CI: 0.915, 0.982), and between the Sutrop index and the sample proportions was 0.86 (average Pearson 0.88 95%CI: 0.753, 0.943). Thus, the three measures were highly correlated in 28 examples that varied in content, number of items, and sample sizevalidating the measurement of a single construct.
To test whether the most salient ideas and themes were captured in smaller samples or with limited probing, we used the sample proportions to estimate item salience and compared the set of most salient items across sample sizes and across more and less probing. Specifically, we defined a set of salient items for each example as those mentioned by 20% or more in the sample of size 20 (because all examples had at least 20) with full-listing (because domains were more detailed). We compared the set of salient items with the set of items obtained at smaller sample sizes and with fewer responses per person.
The set size for salient items (prevalence 20%) was not related to overall domain size, but was an independent characteristic of each domain and whether there were core or prototypical items with higher salience. Most domains had about two dozen items mentioned by 20% or more of the original listing sample (n = 20), but some domains had only a half dozen or fewer items (GoodLeader, GoodTeam2Player, GoodTeam3). With full listing, 26 of 28 examples captured more than 95% of the salient ideas in the first 10 interviews: 18 examples captured 100%, eight examples captured 9599%, one example captured 91%, and one captured 80% ( ). With a maximum of three responses per person, about two-thirds of the salient items (68%) were captured with 20 interviews and about half of the items (53%) were captured in the first 10 interviews. With a sample size of 20, a greater number of responses per person resulted in approximately 50% more items than with three responses per person. Extensive probing resulted in a greater capture of salient items even with smaller sample sizes.
The strict notion of complete saturation as the point where few or no new ideas are observed is not a useful concept to guide sample size decisions, because it is sensitive to domain size and the amount of information contributed by each respondent. Larger sample sizes are necessary to reach saturation for large domains and it is difficult to know, when starting a study, just how large the domain or set of ideas will be. Also, when respondents only provide a few responses or codes per person, saturation may be reached quickly. So, if complete thematic saturation is observed, it is difficult to know whether the domain is small or whether the interviewer did only minimal probing.
Rather than attempting to reach complete saturation with an incremental sampling plan, a more productive focus might be on gaining more depth with probing and seeking the most salient ideas. Rarely do we need all the ideas and themes, rather we tend to be looking for important or salient ideas. A greater number of responses per person resulted in the capture of a greater number of salient items. With exhaustive listing, the first 10 interviews obtained 95% of the salient ideas (defined here as item prevalence of 0.20 or more), while only 53% of those ideas were obtained in 10 interviews with three or fewer responses per person.
We used a simple statistical model to predict the number of new items added by each additional person and found that complete saturation was not a helpful concept for free-lists, as the median sample size was 75 to get fewer than one new idea per person. It is important to note that we assumed that interviews were in a random order or were in the order that the interviews were conducted and were not reordered to any kind of optimum. The reordering of respondents to maximally fit a saturation curve may make it appear that saturation has been reached at a smaller sample size [31].
Most of the examples examined in this study needed sample sizes larger than most qualitative researchers use to reach saturation. Masons [6] review of 298 PhD dissertations in the United Kingdom, all based on qualitative data, found a mean sample size of 27 (range 195). Here, few of the examples reached saturation with less than four dozen interviews. Even with large sample sizes, some domains may continue to add new items. For very large domains, an incremental sampling strategy may lead to dozens and dozens of interviews and still not reach complete saturation. The problem is that most domains have very long tails in the distribution of observed items, with many items mentioned by only one or two people. A more liberal definition of complete saturation (allowing up to two new items per person) allowed for saturation to occur at smaller sample sizes, but saturation still did not occur until a median sample size of 50.
In the examples we studied, most domains were large and domain size affected when saturation occurred. Unfortunately, there did not seem to be a good or simple way at the outset to tell if a domain would be large or small. Most domains were much larger than expected, even on simple topics. Domain size varied by substantive content, sample, and degree of heterogeneity in salience. Domain size and saturation were sample dependent, as the holiday examples showed. Also, domain size estimates did not mean that there are only 73 fruits, rather the pattern of naming fruitsfor this particular sampleindicated a set size of 73.
It was impossible to know, when starting, if a topic or domain was small and would require 15 interviews to reach saturation or if the domain was large and would require more than 100 interviews to reach saturation. Although eight of the examples had sample sizes of 5099, sample sizes in qualitative studies are rarely that large. Estimates of domain size were even larger when models incorporated item heterogeneity (salience). The Fruit example had an estimated domain size of 53 without item heterogeneity, but 73 with item heterogeneity. The estimated size of the Fabric domain increased from 210 to 753 when item heterogeneity was included.
The number of responses per person affected both saturation and the number of obtained items. A greater number of responses per person resulted in a greater yield of domain items. The bits of information obtained in a sample can be approximated by the product of the average number of responses per person (list length) and the number of people in a sample. However, doubling the sample size did not necessarily double the unique items obtained because of item salience and sampling variability. When only a few items are obtained from each person, only the most salient items tend to be provided by each person and fewer items are obtained overall.
Brewer [29] explored the effect of probing or prompting on interview yield. Brewer examined the use of a few simple prompts: simply asking for more responses, providing alphabetical cues, or repeating the last response(s) and asking again for more information. Semantic cueing, repeating prior responses and asking for more information, increased the yield by approximately 50%. The results here indicated a similar pattern. When more information was elicited per person, about 50% more domain items were retrieved than when people provided a maximum of three responses.
Interviewing to obtain multiple responses also affects saturation. With few responses per person, complete saturation was reached rapidly. Without extensive interview probing, investigators may reach saturation quickly and assume they have a sample sufficient to retrieve most of the domain items. Unfortunately, different degrees of salience among items may cause strong effects for respondents to repeat similar ideasthe most salient ideaswithout elaborating on less salient or less prevalent ideas, resulting in a set of only the ideas with the very highest salience. If an investigator wishes to obtain most of the ideas that are relevant in a domain, a small sample with extensive probing (listing) will prove much more productive than a large sample with casual or no probing.
Recently, Galvin [21] and Fugard and Potts [22] framed sample size estimation for qualitative interviewing in terms of binomial probabilities. However, results for the 28 examples with multiple responses per person suggest that this may not be appropriate because of the interdependencies among items due to salience. The capturerecapture analysis indicated that none of the 28 examples fit the binomial distribution. Framing the sample size problem in terms that a specific idea or theme will or will not appear in a set of interviews may facilitate thinking about sample size, but such estimates may be misleading.
If a binomial distribution is assumed, sample size can be estimated from the prevalence of an idea in the population, from how confident you want to be in obtaining these ideas, and from how many times you would like these ideas to minimally appear across participants in your interviews. A binomial estimate assumes independence (no difference in salience across items) and predicts that if an idea or theme actually occurs in 20% of the population, there is a 90% or higher likelihood of obtaining those themes at least once in 11 interviews and a 95% likelihood in 14 interviews. In contrast, our results indicated that the heterogeneity in salience across items causes these estimates to underestimate the necessary sample size as items with 20% prevalence were captured in 10 interviews in only 64% of the samples with full listing and in only 4% (one) of samples with three or fewer responses.
Lowe et al. [25] also found that items were not independent and that binomial estimates significantly underestimated sample size. They proposed sample size estimation from the desired proportion of items at a given average prevalence. Their formula predicts that 36 interviews would be necessary to capture 90% of items with an average prevalence of 0.20, regardless of degree of heterogeneity in salience, domain size, or amount of information provided per respondent. Although they included a parameter for non-independence, their model does not seem to be accurate for cases with limited responses or for large domains.
Warning: Super-heated smoke plumes can cause artifacts in nighttime detections. See: Super heated smoke plumes cause data artifacts in nighttime detections (July 17, ). Please also see: Are there ever false detections or data artifacts?
What active fire data and imagery are available in FIRMS Global?
*Additional higher spatial and temporal resolution data and imagery are available for the USA and Canada. See the FIRMS US/Canada table.
1FDC: Fire Detection and Characterization algorithm
2The pixel size systematically grows from sub-satellite towards the edge of the disk
3FRPPIXEL: Fire Radiative Power algorithm
4Latency refers to the estimated time from satellite observation to availability in FIRMS
5Currently curated for FIRMS US/Canada
See Acronyms section for definitions.
Where can I find the MODIS and VIIRS Active Fire User guides?
For the most current information, use the Active Fire User Guides:
What are the key differences between Standard quality and URT/RT/NRT fire data?
Standard (science quality) data products are an internally consistent, well-calibrated record of Earths geophysical properties to support science. The ultra real-time (URT), real-time (RT), and near real-time (NRT) fire products are generated more rapidly than standard processing allows to meet the needs of the applications community. To facilitate this, a number of changes have been made to the standard processing approach: for the URT and RT, direct broadcast data are used (see Questions about Real-Time (RT) and Ultra Real-Time (URT) active fire data) and for NRT, data downlinked from the satellite are sorted, processed, and delivered in an expedited manner (as Session-based Production Data Sets).
One key difference between the MODIS/Aqua NRT and MODIS/Aqua Standard (science quality) Fire Products is the accuracy of the fire locations (positions or geolocation). Unlike MODIS/Terra, VIIRS/Suomi NPP, VIIRS/NOAA-20 and VIIRS/NOAA-21, the position of the Aqua satellite is not as well known when the NRT Fire Product is produced. Most of the time, the additional error introduced in the reported fire location of the MODIS and VIIRS NRT products is small (< 100 m), but in some situations this position error may be large (several kilometers). In particular, the larger errors may occur after spacecraft maneuvers and during space weather events. The NRT Fire location accuracy is also degraded for MODIS/Terra, VIIRS/Suomi NPP, VIIRS/NOAA-20 and VIIRS/NOAA-21 after spacecraft maneuvers, but the time period is typically shorter than for MODIS/Aqua (< 2 hours for MODIS/Terra, VIIRS/Suomi NPP, and VIIRS/NOAA-20/21 vs. up to 12 hours for MODIS/Aqua). When the Standard Fire products are later processed, the best available satellite position data is used and the quality assurance team removes data degraded by spacecraft maneuvers. In addition, the Standard Fire products are typically later reprocessed at a later and the fire positions may be even more accurate in the reprocessed products.
The MODIS and VIIRS active fire and thermal anomalies products are Level 2 products. Unlike many other Level 2 products, the fire product does not make use of ancillary data, and so their codes are identical to the ones used in standard operations. However, please note:
What does a fire detection mean on the ground?
Satellites take a snapshot of events as they pass over Earth. Each hotspot/active fire detection represents the center of a pixel flagged as containing one or more fires or other thermal anomalies (such as volcanoes). For MODIS the pixel is approximately 1 km and for VIIRS the pixel is approximately 375 m. The location is the center point of the pixel (not necessarily the coordinates of the actual fire). The actual pixel size varies with the scan and track (see: What does scan and track mean?). The fire is often less than the size of the pixel (see: What size fires can be detected?). We are not able to determine the exact location or size of the fire; what we do know is that at least one fire is located within the flagged pixel. Sometimes you will see several active fires in a line. This generally represents a fire front.
Image
Image
What are the different sources of fire data in FIRMS?
Please note: A real-time (RT) and ultra real-time (URT) version of both MODIS and VIIRS active fire data are available for the US and Canada within minutes of satellite observation. See What are the real-time (RT) and ultra real-time (URT) data in FIRMS?.
How often are the active fire data acquired?
The MODIS instruments aboard the Terra and Aqua EOS satellites acquire data continuously, providing global coverage every 1-2 days. The MODIS instrument swath is 2,330 km. Terra passes over the equator at approximately 10:30 a.m. and 10:30 p.m., Mean Local Time (MLT), each day; Aqua passes over the equator at approximately 1:30 p.m. and 1:30 a.m., MLT. There are at least 4 daily MODIS observations for almost every area on the equator, with the number of observations increasing (due to overlapping orbits) closer to the poles.
The VIIRS instruments aboard the Suomi NPP and NOAA-20/21 satellites also acquire data continuously. The 3,040 km VIIRS swath enables 15% image overlap between consecutive orbits at the equator, thereby providing full global coverage every 12 hours. Suomi NPP passes over the equator at approximately 1:30 p.m. and 1:30 a.m., MLT; NOAA-20/21 passes over the equator at approximately 12:40 p.m. and 12:40 a.m., operating about 50 minutes ahead of Suomi NPP. Thanks to its polar orbit, mid-latitudes will experience 3-4 looks a day.
FIRMS is updated regularly as new data are ingested, however it is important to note that the data are derived from instruments onboard polar orbiting satellites (MODIS aboard Terra and Aqua and VIIRS aboard Suomi NPP and NOAA-20/21), and generally these have 2 overpasses a day (more towards the poles - see How often are the active fire data acquired?). FIRMS makes the NRT data available within 3 hours of a satellite observation (on a best effort basis), the RT data are available within 30 minutes of overpass and the URT within 5 minutes. See also What time does the satellite pass over my area?
FIRMS Fire maps are updated every 5 minutes, and the SHP, KML and CSV files (available for the last 24, 48 and 7 days) are updated every 60 minutes. The first place data are posted is on the updated daily fires text files which can be accessed via HTTPS: https://nrt3.modaps.eosdis.nasa.gov/archive/FIRMS. As explained here (https://firms.modaps.eosdis.nasa.gov/active_fire/) you will need an Earthdata Login to download data from this folder. FIRMS web services are generated on the fly and use the most recently processed data.
What time does the satellite pass over my area?
In FIRMS Fire Map and NASA Worldview you can overlay the daily orbit tracks to see satellite overpass times for your area of interest based on the currently selected date. For more precise information regarding anticipated satellite observations, please use the NASA Ocean Color Overpass Predictor. It provides a report of observations conducted by selected satellite sensors for a user-defined location and period of time, including the anticipated time and number of daily/nightly observations, and key satellite view and solar geometry parameters.
MODIS data are available from November (for Terra) and from July (for Aqua) to the present. VIIRS 375 m data are currently available from 20 January to the present (for Suomi NPP) and 1 January to the present (for NOAA-20). VIIRS 375 m (NOAA-21) data are currently available from 17 January to the present.
How are fires detected by satellite?
Fire detection is performed using a contextual algorithm that exploits the strong emission of mid-infrared radiation from fires. The NASA MODIS algorithm examines each pixel of the MODIS swath, and ultimately assigns to each pixel one of the following classes: missing data, cloud, water, non-fire, fire, or unknown. More information can be found in: Giglio, L., Descloitres, J., Justice, C. O., & Kaufman, Y. (). An enhanced contextual fire detection algorithm for MODIS. Remote Sensing of Environment, 87: 273-282. doi:10./S-(03)-6
In keeping with MODIS, the VIIRS algorithm is a hybrid thresholding and contextual algorithm using radiometric signals from 4 micron and 11 micron bands (M13 and M15, respectively) and additional bands and a suite of tests for internal cloud mask and rejection of false alarms. The product primarily contains latitude and longitude data for those pixels classified as thermal anomalies. More information can be found in: Schroeder, W., Oliva, P., Giglio, L., & Csiszar, I. A. (). The New VIIRS 375m active fire detection data product: algorithm description and initial assessment. Remote Sensing of Environment, 143: 85-96. doi:10./j.rse..12.008
Why was a particular fire not detected?
There are several reasons why MODIS or VIIRS may not have detected a fire: The fire may have started and ended between satellite observations; cloud cover, heavy smoke, or tree canopy may completely obscure a fire; occasionally the instruments are inoperable and can observe nothing during these times (see data outages and known issues for MODIS and VIIRS); the fire may have been too small or too cool to be detected. The VIIRS 375 m active fire product provides a greater response over fires of relatively small areas due to its higher spatial resolution and it has improved nighttime performance. For more information on the minimum size of fire that can be detected using MODIS data see "What is the smallest fire size that can be detected?" and "How do I know if a fire detection was missed due to cloud or missing data?"
What are the attributes / fields of the active fire data?
I only see fire data available for the last 7 days on your website. How can I get older data?
Data for the last 2 months can be downloaded as text (TXT) files, older data can be obtained through the MODIS Fire Archive Download Tool or directly from the MODIS Fire Team Science Computing Facility at the University of Maryland. MODIS data are available from November (for Terra) and from July (for Aqua) to the present. VIIRS 375 m data are available from 20 January to the present (for Suomi NPP) and 1 January to the present (For NOAA-20).
What projection are the shapefiles in?
The shapefiles are in the Geographic WGS84 projection.
What is the detection confidence?
The confidence value was added to help users gauge the quality of individual fire pixels included in the Level 2 fire product. The confidence field should be used with caution as it is likely that it will vary in terms of meaning in different parts of the world. Nevertheless, some of our end users have found such a field to be useful in excluding false positive occurrences of fire. The confidence values are attributed differently for MODIS and VIIRS.
For MODIS, the confidence value ranges from 0% to 100% and can be used to assign one of the three fire classes (low-confidence fire, nominal-confidence fire, or high-confidence fire) to all fire pixels. For some uses or applications, errors of commission (or false alarms) are particularly undesirable, and for these applications a user may consider trading a lower detection rate to gain a lower false alarm rate. Conversely, for other applications missing any fire might be especially undesirable, and a user may be willing to tolerate a higher false alarm rate to ensure that fewer true fires are missed. Users requiring fewer false alarms may wish to retain only nominal- and high-confidence fire pixels and treat low-confidence fire pixels as non-fire, land pixels. Users requiring maximum fire detectability who are able to tolerate a higher incidence of false alarms should consider all three classes of fire pixels.
For VIIRS, the confidence values are set to low, nominal, and high; they are based on a collection of intermediate algorithm quantities used in the detection process and are intended to help users gauge the quality of individual hotspot/fire pixels. Low confidence daytime fire pixels are typically associated with areas of Sun glint and lower relative temperature anomaly (<15 K) in the mid-infrared channel I4. Nominal confidence pixels are those that are free of potential Sun glint contamination during the day and marked by strong (>15 K) temperature anomaly in either day or nighttime data. High confidence fire pixels are associated with day or nighttime saturated pixels.
While the confidence value may not be very helpful for all users or applications, there is no way to establish an optimal cutoff a priori. Users have to adopt an empirical approachwhat threshold works best for what I'm trying to do?
What are MODIS and VIIRS collections?
Reprocessing of the entire MODIS and VIIRS data archive is periodically performed to incorporate better calibration, algorithm refinements, and improved upstream products into all MODIS or VIIRS products. The updated data archive resulting from each reprocessing is referred to as a collection. Later collections supersede all earlier collections.
For Terra MODIS, Collection 1 consists of the first products generated following launch. Terra MODIS data were reprocessed for the first time in June to produce Collection 3. Note that this first reprocessing was numbered Collection 3, rather than Collection 2, as one would expect. Collection 3 was also the first produced for the Aqua MODIS products. Collection 4 reprocessing was initiated in December for Terra MODIS and somewhat later for the Aqua MODIS. Collection 5 began reprocessing in early . Improvements in Collection 5 included adding the Fire Radiative Power value to the fire detections and refining the detection confidence to more accurately identify questionable active fire pixels. Collection 6 reprocessing began in . In terms of fire detections, the key differences between Collection 5 and Collection 6 were that: Collection 6 extends processing to oceans and other large bodies, including detection of off-shore gas flaring; there was a reduction in false alarms in the Amazon caused by small forest clearings. Collection 6.1 has been available since April and is the latest collection available. In terms of fire detections, there are no algorithm changes; the update is from changes and enhancements to the calibration approach used in the generation of the Terra and Aqua MODIS Level 1B products. For further details on C61 calibration changes and other changes, the user is encouraged to refer to the pdf summarizing Collection 6.1 specific changes.
For the most up to date information, see the MODIS Collection 6/6.1 Active Fire Product User's Guide (updated May )
Are there any restrictions on using data or imagery from FIRMS?
NASA supports an open data policy and we encourage the appropriate use of data and graphics from FIRMS; when doing so, please take a moment to make sure you get the correct citationsee the Citation Policy. Please also read the LANCE Disclaimer and the About FIRMS web page.
What is UTC?
UTC is known as "Universal Coordinated Time, Coordinated Universal Time or Zulu Time. It is a worldwide scientific standard of timekeeping. Its reference point is Greenwich, England: when it is midnight there on Earth's prime meridian, it is midnight (00:00:00.). UTC can be converted to local standard time for a particular time zone by adding or subtracting the time difference between UTC and that time zone. Please see this World Time Zone Map to visualize time zones boundaries and differences between UTC and local standard time for the respective time zones. Please note when daylight savings time is applied in a given time one hour should be subtracted from the UTC/time zone difference.
Why do I sometimes see linear patterns of large numbers of active fire detections?
NASA and NOAA routinely conduct on-orbit mission calibration maneuvers for MODIS, VIIRS, etc. to maintain consistent geometric and radiometric performance for production of science data. Primarily for radiometric calibration purposes, on a near monthly basis the sensors typical perspective of the Earths is changed to view the Moon. This is conducted by physically rolling the satellite/spacecraft over in orbit for a short period of time to observe the Moon. The illumination of the Moons surface by the Sun during these maneuvers is used to calibrate the reflective solar bands of the sensor. Upon completion of these maneuvers the satellite rolls back to point the sensor at Earth. At that instant, it is not uncommon for anomalous observation data to be collected by the reflective and thermal bands. This situation can result in linear patterns of large numbers of anomalous active fire detection data within a MODIS or VIIRS swath as show in the example below.
Image
NASA FIRMS makes every effort to filter and remove these anomalous detections from the data stream in real-time when they are detected. However, on rare occasions it is possible for some of these detections to temporarily appear in the FIRMS data viewer and datasets before they are removed.
Back to top
Are there ever false fire detections or data artifacts?
To minimize the occurrence of false alarms over bright/reflective surfaces (e.g., metallic factory rooftops), the VIIRS and MODIS algorithms use slightly more conservative tests to avoid the effects of Sun glint.
Super-Heated Smoke Plumes
A few VIIRS active fire pixels are occasionally located outside the reported perimeter of a large wildfire. The parallax effect causes the tall/superheated plume detection pixel(s) to be displaced laterally when projected onto the surface. Displaced pixels will be located on the fire perimeters side further away from the image center and closer to the swaths edge. If this happens, users may need to look for additional clues to see if it is a data artifact. There have been a few instances when tall, superheated plumes carrying large volumes of hot material into the air are formed over large and intense wildfires; when these occur the VIIRS instrument detects the surface fire along with part of the plume and categorizes them as active fires. Those occurrences typically share the following set of conditions:
(i) Nighttime detection. This is the period during which the VIIRS active fire product is particularly responsive to heat sources thereby favoring plume detection. However, this can also occur during the daytime with;
(ii) Very large wildfires undergoing explosive growth and accompanied by rapid/vertically elongated plume development. Enough hot material must entrain the plume creating a distinguishable thermal signal (i.e., one that significantly exceeds the fire-free surface background);
(iii) High scan angle. This is what will ultimately produce the detections extending beyond the actual fire perimeter.
If those conditions apply, look for alternative observations (previous/next observation) acquired closer to nadir, and try to prioritize the use of the fire detection data accordingly. Unfortunately, the VIIRS active fire algorithm isnt currently able to distinguish nighttime surface fire pixels from the isolated plume detections due to strong similarities between their radiometric signatures.
Related content:
How do I know if a fire detection was missed due to cloud or missing data?
An indication of cloud cover or missing data can be obtained by viewing the co-incident MODIS or VIIRS imagery in FIRMS Fire Map or Worldview. To take cloud and missing data in to account, it may be more appropriate to use one of the 1 km Level 3 or Climate Modeling Grid (CMG) fire products (see the discussion in the MODIS Fire User Guide).
How appropriate are the hotspot/fire locations for my research?
The MODIS and VIIRS fire locations are good for determining the location of active fires, providing information on the spatial and temporal distribution of fires, and comparing data between years. The active fire pixel locations may not always be the most appropriate source of fire related information, and, the data do not provide any information on cloud cover or missing data. Depending on the analysis you are performing, misleading or even incorrect results may result, if the other types of pixels are ignored. In some cases, it is more appropriate to use one of the Level 3 or CMG fire products. For more discussion on this, see FAQ "What is the Climate Modeling Grid (CMD) fire product? and "What is the difference between NRT and Standard quality fire data?"
What is the Climate Modeling Grid (CMG) fire product?
The CMG fire products are gridded statistical summaries of fire pixel information intended for use in regional and global modeling. The products are currently generated at 0.5 degree spatial resolution for time periods of 1 calendar month (MOD14CMH/MYD14CMH) and 8 days (MOD14C8H/MYD14C8H). Higher resolution 0.25 degree CMG fire products will eventually be produced as well. More information can be found in the MODIS Collection 6 Active Fire Product User's Guide.
Can active fires be detected below the forest canopy?
The likelihood of detecting a fire beneath the tree canopy is unknown, but likely to be very low. Understory fires are typically small, and with the tree canopy obstructing the view of the fire, detection will be very unlikely.
Are all active fire hotspots vegetation fires?
No. An active fire represents the center of a pixel flagged as containing 1 or more actively burning hotspots /fires. In most cases hotspots/fires are vegetation fires, but sometimes a hotspot/fire is a volcanic eruption or the flare from a gas well. Different types of thermal anomalies are not currently attributed in the NRT MODIS or VIIRS data.
Can I estimate burned area from active fire data?
It is not recommended to use active fire locations to estimate burned area as determining burned area to an acceptable degree of accuracy is not possible due to nontrivial spatial and temporal sampling issues. For some applications, however, acceptable accuracy can be achieved, although the effective area burned per fire pixel is not simply a constant, but rather varies with respect to several different vegetation and fire-related variables. See Giglio et al. () for more information. See "Where can I get the MODIS burned area product?"
Most recently, Harmonized Landsat Sentinel (HLS) false color composite imagery was included in FIRMS and these data can be helpful in estimating burned area (see FIRMS Blog post on adding HLS).
Can you use the active fire data for detecting volcanoes or volcanic eruptions?
While both the MODIS and VIIRS algorithms routinely detect active volcanoes, the active fire products have not been validated against independent data for their ability to detect volcanoes. There is a separate near-real time MODIS product (MODVOLC) specifically for volcanoessee http://modis.higp.hawaii.edu/.
What is FRP and what factors should be considered across sensors?
Fire Radiative Power (FRP) is the rate of radiative energy emission per time unit from all fires within a pixel. FRP has been extensively used as a proxy of fire intensity to characterize fire types and is related to the rate of biomass combustion and rate of emissions. FRP estimates are provided for active fire data from MODIS, VIIRS and geostationary sensors used in FIRMS. Sensor spatial resolution, saturation temperature of thermal bands, satellite view zenith angle, time of satellite observation, etc. can affect the accuracy and variability of FRP estimates. Users should consider these factors when comparing FRP information from a specific sensor or multiple sensors.
Back to top
The Visible Infrared Imaging Radiometer Suite (VIIRS) is an instrument aboard the Suomi National Polar-orbiting Partnership (Suomi NPP) and NOAA-20/21 weather satellites. VIIRS was launched on Suomi NPP on October 28, , and again on NOAA-20 (formally known as Joint Polar Satellite System-1 [JPSS-1]) on November 18, . The VIIRS instrument was launched on JPSS-2 (NOAA-21) on November 10, and has joined NOAA-20 and Suomi NPP in the same orbit.
If you wish to view the VIIRS reflectance imagery that corresponds to the active fire detections, you can display this in FIRMS Fire Map or Worldview.
What is the VIIRS 375 m Active Fire Product?
The VIIRS 375 m active fire product is the latest product to be added to FIRMS. It provides data from the Visible Infrared Imaging Radiometer Suite (VIIRS) sensor aboard the joint NASA/NOAA Suomi National Polar-orbiting Partnership (Suomi NPP), NOAA-20 and NOAA-21 satellites. The 375 m data complements MODIS fire detections, show good agreement in hotspot detection, but the improved spatial resolution of the 375 m data provides greater response over fires of relatively small area, and improved mapping of large fire perimeters. The 375 m data also has improved nighttime performance. Consequently, these data are well suited for use in support of fire management (e.g., near real-time alert systems), as well as other science applications requiring improved fire mapping fidelity.
The VIIRS NRT 375 m active fire products are from: Suomi NPP (VNP14IMGTDL_NRT);NOAA-20, formally known as JPSS-1 (VJ114IMGTDL_NRT);and NOAA-21, formally known as JPSS-2 (VJ214IMGTDL_NRT).
What is the temporal frequency of the VIIRS 375 m fire data?
The VIIRS instrument aboard the Suomi NPP and NOAA-20/21 satellites acquire data continuously. The 3,040 km VIIRS swath enables 15% image overlap between consecutive orbits at the equator, thereby providing full global coverage every 12 hours. Suomi NPP has a nominal (equator-crossing) observation times at 1:30 p.m. and 1:30 a.m., Mean Local Time (MLT); NOAA-21 is flying half an orbit, or 50 minutes, ahead of NOAA-20, with Suomi NPP between them. The mid-latitudes experience 3-4 looks a day as these are polar-orbiting satellite.
What is the spatial resolution of the data?
The 375 m data describe the nominal resolution after native data are spatially aggregated (see figure below).
Image
Spatial resolution of VIIRS imagery data as a function of scan angle. The three distinct regions represent unique data aggregation zones extending from nadir to the edge of swath.The aggregation scheme changes across three distinct image regions. In the first region (nadir to 31.59° scan angle), three native pixels are aggregated in the along scan (cross-track) direction to form one data sample in the Level 1 image. In the second region (31.59° to 44.68° scan angle), two native pixels are aggregated to form one data sample. Finally in the third and last region (44.68° to 56.06° - edge of swath) one native pixel will result in one data sample. All five 375 m channels are aggregated onboard the spacecraft before the data are transmitted to the ground stations, whereas a subset of the VIIRS 750 m data (dual-gain channels only) are aggregated on the ground.
What is the main difference between the VIIRS 375 m and 750 m active fire data?
The two data products use similar methodologies to detect active fire pixels although differences in the spectral characteristics of the VIIRS channels used in each case led to unique algorithms. Because of its improved spatial resolution, the 375 m algorithm will tend to detect more fire pixels compared to the 750 m data set. That difference is particularly pronounced during the nighttime part of the orbit when the occurrence of smaller/cooler fires will favor the 375 m product.
Will the VIIRS 375 m fire detection algorithm always outperform the 750 m one?
Generally, the higher resolution product will achieve a higher probability of fire detection in both day and nighttime scenes. However, areas subject to strong solar reflectance associated with sun glint, such as metallic rooftops or highly reflective desert, could see a few 750 m fire pixels without a corresponding 375 m fire detection. This is attributed to the relatively shorter wavelength of the 375 m mid-infrared channel used in the fire algorithm, which will experience greater influence of the solar component. To minimize the associated consequences, namely the occurrence of false alarms over bright/reflective surfaces (e.g., metallic factory rooftops), the 375 m algorithm uses slightly more conservative tests to avoid the effects of Sun glint over those areas.
Are those few isolated fire pixels in the middle of the South Atlantic Ocean real?
Those occurrences are typically associated with spurious fire detections due to the South Atlantic Magnetic Anomaly. The 375 m active fire algorithm contains a specific filter to flag those occurrences as low confidence detections. However, in some cases (average 2-3 pixels every night) the spurious signal generated in the input Level 1 data is confused for a regular fire pixel and therefore assigned a nominal confidence flag. Note that verified true positives can be found over South Atlantic Ocean waters along the southeast coast of Brazil and the west coast of Africa where oil rigs normally operate.
Are the VIIRS 375 m (and 750 m) active fire data science-ready?
The VIIRS active fire data have been extensively tested since routine production of the missions data record started January 19, . Numerous bad scan episodes (i.e., pixel clusters containing spurious radiances extending across the swath) were found in the Level 1 input data during the initial 18-24 months of the time series (Csiszar, et al., ). Those anomalies were gradually addressed by the VIIRS science team and their occurrence has dropped to virtually zero with the implementation of revised Level 1 data processing packages in . Initial assessment of both VIIRS 375 m and 750 m was implemented over a few experimental sites indicating consistent fire detection and characterization performance. Additional data comparison analyses were implemented using near coincident Aqua/MODIS and TET-1 (German Aerospace Center) active fire data, which again showed consistent performance of the VIIRS active fire products across different observation conditions. Consequently, we consider the current data of good enough quality for use in fire management applications and scientific studies.
Where do I go for more information?
Schroeder, W., Oliva, P., Giglio, L., & Csiszar, I. A. (). The New VIIRS 375m active fire detection data product: Algorithm description and initial assessment. Remote Sensing of Environment, 143: 8596. doi:10./j.rse..12.008
Back to top
The Moderate Resolution Imaging Spectroradiometer (MODIS) is a payload imaging sensor that was launched into Earth orbit by NASA in aboard the Terra (EOS AM) satellite and in aboard the Aqua (EOS PM) satellite.
MODIS reflectance imagery that corresponds to the active fire detections, can be viewed in FIRMS Global, FIRMS US/Canada and Worldview.
What does a MODIS active fire detection mean on the ground?
Satellites take a snapshot of events as they pass over earth. Each hotspot/active fire detection represents the center of a 1 km (approx.) pixel flagged as containing one or more fires or other thermal anomalies (such as volcanoes). The location is the center point of the pixel (not necessarily the coordinates of the actual fire). The actual pixel size varies with the scan and track (see: What does scan and track mean?). The fire is often less than 1 km in size (see: What size fires can be detected?). We are not able to determine the exact fire size; what we do know is that at least one fire is located within that 1 km pixel. Sometimes you will see several active fires in a line. This generally represents a fire front.
Image
Are there any missing MODIS fire data?
Terra was launched 18 December and Aqua was launched 4 May . High quality active fire observations are available from the Terra satellite starting November and from the Aqua satellite starting 4 July onwards.
In the fire data archive, there are several days where data were not collected and days with lower than usual fire counts due to reasons such as sensor outage. These include, but are not limited to: 15 April , 15 June to 3 July , and 19 to 28 March .
What does scan and track mean?
Scan and Track values are included in the attributes (see: What are the attributes / fields of the active fire data?). The scan value represents the spatial-resolution in the East-West direction of the scan and the track value represents the North-South spatial resolution of the scan. It should be noted that the pixel size is not always 1 km across the scan track. The pixels at the Eastern and the Western edges of the scan are bigger than 1 km. It is 1 km only along the nadir (exact vertical from the satellite). Thus, the values shown for scan and track represent the actual spatial resolution of the scanned pixel.
What is the brightness temperature?
The brightness temperature of a fire pixel is measured (in Kelvin) using the MODIS channels 21/22 and channel 31. Brightness temperature is actually a measure of the photons at a particular wavelength received by the spacecraft, but presented in units of temperature.
Where can I get the MODIS burned area product?
MODIS Active Fire & Burned Area Products web site and the MODIS Burned Area User Guide (Updated November ) and the MODIS Collection 6 and 6.1 Active Fire Product User Guide.
What size fires can be detected?
In any given scene the minimum detectable fire size is a function of many different variables (scan angle, biome, Sun position, land surface temperature, cloud cover, amount of smoke and wind direction, etc.), so the precise value will vary slightly with these conditions. MODIS routinely detects both flaming and smoldering fires m2 in size. Under very good observing conditions (e.g., near nadir, little or no smoke, relatively homogeneous land surface, etc.) flaming fires one tenth this size can be detected. Under pristine (and extremely rare) observing conditions even smaller flaming fires 50 m2 can be detected. It is not recommended to estimate burned area from the active fire data, see: Can I estimate burned area using the active fire data?.
Unlike most contextual fire detection algorithms designed for satellite sensors that were never intended for fire monitoring (e.g., AVHRR, VIIRS, ATSR), there is no upper limit to the largest and/or hottest fire that can be detected with MODIS.
Image
The diagram shows the day and night relationship of fire size and fire temperature, in different biomes, to the probability of being detected by MODIS (Giglio, et al. ()).
What validation of the MODIS active fire products has been performed?
Validation of the Terra MODIS Fire Product has primarily been performed using coincident observations from the Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER); see the MODIS Land Team Validation page, and publications by Csiszar, et al. () and two publications from Morisette, et al. () for details. A very brief discussion of the general validation procedure, with some preliminary results, can be found in the Justice, et al. () publication.
Where can I get more information on the MODIS Fire Products?
For more information on the active fire product and other MODIS fire products, please refer to the papers by Louis Giglio found here as well as the MODIS Active Fire Burned Area Products website and the MODIS Burned Area User Guide (Updated November ) and the MODIS Collection 6 Active Fire Product User Guide (Updated December ).
Back to top
What are the real-time (RT) and ultra real-time (URT) data in FIRMS?
NASA EOSDIS defines real-time (RT) data as data that are available to the end user within one hour of the satellite fly-over (see What is Data Latency? and What time does the satellite pass over my area?). Ultra real-time (URT) data are available much quickerwithin minutes. Within FIRMS, the URT data are available within a minute of satellite overpass for select direct readout stations in the U.S. and Canada.
What is "direct broadcast"?
Are you interested in learning more about Cropped and Flamed Setts? Contact us today to secure an expert consultation!
Direct broadcast, in addition to storing collected observation data for later download at designated intervals, some satellites have the capability to immediately broadcast this raw data to the ground. This is known as direct broadcast. As Earth is observed by the satellite, the data are formatted and transmitted to any user below with an antenna and other necessary equipment to receive and process the data. The Terra, Aqua, Suomi NPP, and NOAA-20/21 satellites all have direct broadcast capability.
What is "direct readout"?
The acquisition of freely transmitted direct broadcast satellite data by users with compatible ground receiving equipment is known as direct readout. Data can be collected by direct readout ground stations while they are in direct line of sight to satellites while they are visible above the horizon.
Are there particular differences between URT data, RT data and NRT data?
Ultra Real-Time (URT), Real-Time (RT), and Near Real-Time (NRT) data are all relative forms of low-latency data products made available more quickly than standard processing to support environmental monitoring, response to hazard/disaster events, etc.
NRT data processing frameworks, such as NASA LANCE, leverage operational Level 0 global data streams for several NASA missions and provide Level 2 products within 1-3 hours of satellite overpass. RT and URT data are predicated on the use of raw satellite data collected via direct readout ground stations. These ground stations are operated by various agencies and institutions, located around the globe and provide regional coverage. Ground stations typically process raw direct readout data to Level 0 to Level 1 to Level 2 data 30-60 minutes over satellite overpass and are considered RT data. Recently developed data processing technologies can now process direct readout data from Level 0 to Level 2 data products within 1 minute of satellite observation and are considered URT data.
How are URT active fire detection data integrated and used with NRT active fire detection data in FIRMS?
URT active fire detection data provided by select ground stations are integrated into FIRMS fire maps, KMZs and web map services within minutes of satellite observation. NRT active fire detection data replace the URT data in FIRMS mapping and data products and services as they become available 1-3 hours later. This integration of the data is seamless and not noticeable to the user unless there is a temporary anomaly occurring in either data source (e.g., a gap or artifact in the URT direct readout data stream).
How do I identify the NRT, RT and URT data?
FIRMS Fire Map and Map Services: To view the NRT, RT, and URT data in the FIRMS fire map, select the "Time Based" option in the FIRMS menu on the right. This symbolizes the active fire detections by time since detection. Both RT and URT data will display in the <1 hour category while NRT data will display in the 1-3 hour category or later categories. You can also click on individual active fire detection locations to bring up the attributes table; column X lists whether it is NRT, RT or URT data. The RT and URT fires will have a Version value (e.g., 2.0) followed by RT or URT. NRT fires will have a Version value followed by NRT.
FIRMS KMZs: "Active fire detection footprint and centroid" KMZs symbolize active fire detection by time since detection. Time since detection KMZs symbolizes the active fire detections by time since detection. Both RT and URT data will display in the <1 hour category while NRT data will display in the 1-3 hour category or later categories. RT and URT data are not distinguished in the KMZ attributes.
Are URT data provided for all FIRMS products?
So far, URT active fire detection data are available in FIRMS interactive maps, KMZs, and web services for most of the conterminous U.S. plus S Canada and N Mexico.
Are URT data provided globally?
URT data are currently provided by four (4) select direct readout ground stationsone in Mayaguez, PR (red) and Hampton, VA (black) and two in Madison, WI (blue) (see graphic below). Each ground station can receive direct readout data when satellite orbits intersect their respective station masks. Additional United States ground stations will continue to be integrated with the objective of providing URT coverage for the U.S., Canada, and Central America. Integration of ground stations elsewhere on the globe will be evaluated and determined at a later date.
Image
Current (solid lines) and Planned (dotted lines) Coverage of Ultra Real-Time and Real-Time Active Fire DataWhat is the objective of providing URT data?
Remote sensing observations and derived science products provide a "snapshot" of conditions at the time of satellite observation and are used to assess the situation and inform decisions. Consequently, the value of satellite data used for operational detection and monitoring of wildfires and other hazard/disaster events is inversely proportional to their latency. In other words, the value of the data typically increases with decreasing latency in data availability since satellite observation.
Back to top
How are "geostationary" active fire data different from the other data provided by FIRMS?
Geostationary satellite sensors view the same area of Earths surface at all times while polar-orbiting sensors, such as MODIS and VIIRS, typically view the same area of Earths surface twice daily. Consequently, geostationary satellite sensors can provide repeated observations on a sub-hourly basis, making it possible to detect fires which may not be detected at longer temporal intervals. Geostationary satellites provide data at 10-15 minute intervals, so they can detect more fire events and capture their growth and change. However, the spatial resolution of geostationary satellite data is coarser and therefore less sensitive to small fires.
Where does the geostationary active fire data come from and how can I get more information about them?
Please refer to the following references and the information in the table below:
*See Coverage Figure below
What are the attributes of the geostationary fire data?
The attributes are described in the following table:
Attribute Short Description Long Description Latitude Latitude (decimal degrees) Latitude of pixel center. Note: Pixel size/dimension is affected by satellite view geometry and curvature of Earths surface. Consequently, pixel size becomes more distorted (larger) with increasing distance from the point immediately below the satellite. The reported coordinate is the center of each pixels footprint. See: What is the spatial resolution of the geostationary satellite observations? Longitude Longitude (decimal degrees) Longitude of pixel center. Note: Pixel size/dimension is affected by satellite view geometry and curvature of Earths surface. Consequently, pixel size becomes more distorted (larger) with increasing distance from the point immediately below the satellite. The reported coordinate is the center of each pixels footprint. See: What is the spatial resolution of the geostationary satellite observations? Brightness Brightness Temperature (BT) (kelvins) Middle-infrared channel BT (kelvins)FRP-PIXEL: Value range 0-1. Higher values correspond to higher detection confidences.
FDC-HSC: class values range from 10-15 and 30- 35 and are indicative, but not a measure of fire confidence level. The former sub-group defines instantaneous (first/single) fire detections whereas the latter identifies temporally-filtered cases when two or more co-located detections are observed within a 12-hour interval.
FIRMS filters and displays only:
Long-wave infrared channel brightness temperature (kelvins)
No BT values are output from the FDC algorithm indicated by -999
Him8 = Himawari-8 KCL/IPMA
Met9 = Meteosat-9 LSA SAF
Met11 = Meteosat-11 LSA SAF
G16 = GOES-16 (GOES-East) NOAA FDC
G18 = GOES-18 (GOES-West) NOAA FDC
G16FRP = GOES-16 (GOES-East) KCL/IPMA
G18FRP = GOES-18 (GOES-West) KCL/IPMA
What is the temporal frequency of the geostationary satellite observations and derived active fire data?
What size of fires can be detected with geostationary satellite observations?
GOES - The theoretical minimum detectable fire size is approximately 0.004km2 (4,000m2) at nadir in optimum viewing conditions, burning at 800K (Schmidt et al., ).
Meteosat - The theoretical minimum detectable fire size for actively burning fires is approximately 0. of the area of a 3km SEVIRI pixel at nadir or approximately 0.km2 (900m2) (Schmidt et al., ).
Himawari - Documentation describing the minimum detectable fire size of Himawari AHI using JAXA's Wild Fire product algorithm is currently limited but performance is anticipated to be similar to GOES ABI. Several recent research efforts demonstrate the use of Himawari AHI using customized/adapted fire detection algorithms with varying results.
Please see the following papers:
Schmidt, C., Hoffman, J., Prins, E., and Lindstrom, S. (). GOES-R Advanced Baseline Imager (ABI) Algorithm Theoretical Basis Document For Fire/ Hot Spot Characterization v2.6. NOAA NESDIS STAR, https://www.star.nesdis.noaa.gov/goesr/documents/ATBDs/Baseline/ATBD_GOESR_FIRE_v2.6_Oct.pdf
Wooster, M. J., Roberts, G., Freeborn, P. H., Xu, W., Govaerts, Y., Beeby, R., He, J., Lattanzio, A., Fisher, D., and Mullen, R. (). LSA SAF Meteosat FRP products Part 1: Algorithms, product contents, and analysis. Atmospheric Chemistry and Physics, 15(22): . doi:10./acp-15--
What is the spatial resolution of the geostationary satellite observations?
The thermal bands, used to detect active fires, have a spatial resolution directly below the satellite (known as the sub-satellite point) of 2km for GOES ABI and Himawari AHI, and 3km for Meteosat SEVIRI. However, as already noted, the spatial resolution of the pixel (i.e., the pixel size) systematically grows away from the sub-satellite point towards the edge of the disk (Figures 1-4). Please note each hotspot/active fire detection represents the center of a pixel flagged as containing one or more fires or other thermal anomalies (such as volcanoes).
Image
Please note the pixel size systematically grows away from the sub-satellite point towards the edge of the disk (see Figures 1 - 4).The following figures (Figures 1-4) illustrate how the spatial resolution of the geostationary satellites increases away from the sub-satellite point - please note all measurements are in km2.
Image
Figure 1. Coverage of GOES-18 indicating the variation in pixel area (in km²)Image
Figure 2. Coverage of GOES-16 indicating the variation in pixel area (in km²).Image
Figure 3. Coverage of Meteosat SEVIRI indicating the variation in pixel area (in km²).Image
Figure 4. Coverage of Himawari-8 indicating the variation in pixel area (in km²).Why are the geostationary fire data filtered in FIRMS?
While geostationary active fire data provide a much enhanced temporal frequency compared to active fire data generated from polar-orbiting satellites, active fire data from geostationary satellites can yield spurious detections. These can result from sun glint, bright water, bright deserts/soils, coastlines, and navigation inconsistencies resulting in positioning error, for example. The Meteosat SEVIRI and Himawari AHI data, made available from EUMETSAT and CAM, have a 65-degree view zenith angle cut-off due to the large pixel size at high latitudes.
The following images provide examples of where spurious fire detections from Meteosat SEVIRI have been noted.
Image
This series (1 of 3) of FIRMS screenshots, from August 11, , show fires detected in the Serra da Estrela National Park, Portugal. This first screenshot shows a MODIS/Terra 7,2,1 Corrected Reflectance image with burned area in reddish-brown and an active fire front in orange (note: no active fires have been overlayed).Image
This second image in the series of FIRMS screenshots from August 11, , shows fires detected in the Serra da Estrela National Park, Portugal. This screenshot shows a MODIS/Terra 7,2,1 Corrected Reflectance image, overlain with VIIRS active fire detections shown as red pixels.Image
This third image in the series of FIRMS screenshots from August 11, , shows fires detected in the Serra da Estrela National Park, Portugal. This third screenshot shows a MODIS/Terra 7,2,1 Corrected Reflectance image, overlain with VIIRS active fire detections shown as red pixels and Meteosat-11 SEVIRI FRP pixels shown as orange pixels. Some spurious SEVIRI fire detections are evident.The Filtered Fire Detection and Characterization (FDC-HSC) algorithm used to produce the GOES-16 and GOES-18 active fire data from NOAA, is still under development. To limit false detections, the data shown in FIRMS are filtered to show only fire detections with higher confidence values; these are classes 10 and 30, and 11 and 31. These criteria are driven by the validation analysis (See Hall et al ) and a qualitative review of the current algorithm output. The full list of classes available is outlined in the table below:
Class Description 10 or 30 Good quality (highest confidence) fire detections** 11 or 31 Saturated fire pixel** 12 or 32 Cloud contaminated fire pixel 13 or 33 High probability fire pixels that do not satisfy requirements to be categorized as good quality fire detections 14 or 34 High probability fire pixels that do not satisfy requirements to be categorized as good quality fire detections 15 or 35 Low probability fire pixels that do not satisfy requirements to be categorized as good quality fire detections**indicates the classes shown in FIRMS
Our qualitative assessment indicates some spurious fire detections are showing (see example below). This could be due to sun glint, bright water, bright deserts/soil, coastlines and lake edges etc so users are requested to use the data with caution. If in doubt, we recommend users look for coincident fire detections from MODIS or VIIRS in the vicinity. If they are present, the user can have more confidence in the geostationary data.
Image
In this screenshot from FIRMS, on 15 July , the GOES FDC data from NOAA (shown in pink) along the lake edge are thought to be false positives / spurious fire detections. Users should exercise caution when using the data.What validation of geostationary active fire data has been conducted?
Satellite-derived active fire products are validated with spatially and temporally coincident observations from higher resolution sensors. Landsat and Sentinel-2 are both appropriate sensors for validating geostationary active fire products. Given the Landsat and Sentinel-2 overpass times, any validation efforts will be focused on certain times of the day and, therefore, false positives associated with unique solar angle issues, for example, may be missed.
Please see the following papers:
Hall, J. V., Zhang, R., Schroeder, W., Huang, C., & Giglio, L. (). Validation of GOES-16 ABI and MSG SEVIRI active fire products. International Journal of Applied Earth Observation and Geoinformation, 83, . doi:10./j.jag..
Xu, W., Wooster, M.J., Kaneko, T., He, J., Zhang, T. and Fisher, D., . Major advances in geostationary fire radiative power (FRP) retrieval over Asia and Australia stemming from use of Himawari-8 AHI. Remote Sensing of Environment, 193, pp.138-149. doi:10./j.rse..02.024
Xu, W., Wooster, M.J., He, J. and Zhang, T., . Improvements in high-temporal resolution active fire detection and FRP retrieval over the Americas using GOES-16 ABI with the geostationary Fire Thermal Anomaly (FTA) algorithm. Science of Remote Sensing, 3, p.. doi:10./j.srs..
Wooster, M. J., Roberts, G., Freeborn, P. H., Xu, W., Govaerts, Y., Beeby, R., He, J., Lattanzio, A., Fisher, D., and Mullen, R. (). LSA SAF Meteosat FRP products Part 1: Algorithms, product contents, and analysis. Atmospheric Chemistry and Physics, 15(22): . doi:10./acp-15--
Roberts, G., Wooster, M.J., Xu, W., Freeborn, P.H., Morcrette, J.-J., Jones, L., Benedetti, A., Jianping, H., Fisher, D., Kaiser, J.W. (). LSA SAF Meteosat FRP products - Part 2: Evaluation and demonstration for use in the Copernicus Atmosphere Monitoring Service (CAMS). Atmospheric Chemistry and Physics, 15(22): . doi:10./acp-15--
Can you provide a brief description of the two algorithms used to produce geostationary active fire products from GOES-16 and 18?
FRP-PIXEL Product
The key outputs from the FRP-PIXEL algorithm are active fire pixel detections and fire characteristics including fire radiative power (FRP), FRP uncertainty, active fire temperature, and fire detection confidence levels. The detection algorithm used in the FRP-PIXEL product is based on the principles applied to generate active fire detections within the MODIS Fire Products (Giglio et al., ). The algorithm works on statistics derived primarily from middle-infrared (MIR) and thermal infrared (TIR1) brightness temperature imagery. On a first pass, a series of absolute thresholds applied to these data are used to detect potential fire pixels. These pixels are then further assessed as true or false fire detections based on a series of further contextual tests based on statistics derived from immediate neighboring non-fire background pixels. Further description of the algorithm can be found in (Govaerts et al., ; Wooster et al., ).
The FRP-PIXEL product is available every 15 minutes for the SEVIRI full-disk and every 10 minutes for the AHI and ABI disks. The spatial resolution of the product at the sub-satellite point is 3-km for the SEVIRI data and 2-km for the AHI and ABI, but the pixel size systematically grows with increasing view zenith angle (VZA) from the sub-satellite point towards the edge of the disk (See: What is the spatial resolution of the geostationary satellite observations?).
FDC-HSC Product
The key outputs from the FDC-HSC algorithm include information on the fire pixel locations, sub-pixel fire characteristics )active fire area and temperature), and FRP. The current FDC-HSC algorithm is a multispectral contextual algorithm designed to identify fire pixels through a decision-tree approach comparing them with their neighboring non-fire background pixels. The FDC-HSC algorithm uses a two-step approach. The first step is designed to identify all potential fire pixels while the second step further confirms and then characterizes the sub-pixel fire activity. The FDC-HSC product is available every 10 minutes. The spatial resolution of the product at the sub-satellite point is 2km, and the pixel size systematically grows with increasing VZA from the sub-satellite point towards the edge of the disk.The FDC-HSC product is available every 10 minutes. The spatial resolution of the product at the sub-satellite point is 2km, and the pixel size systematically grows with increasing VZA from the sub-satellite point towards the edge of the disk.
What are the Product Uncertainties and Limitations?
The FRP-PIXEL and FDC-HSC products are subject to several sources of uncertainties due to the characteristics of geostationary satellite full-disk data and associated image processing. Among these limitations:
Back to top
What is the Landsat Fire and Thermal Anomalies Product?
The Landsat Fire and Thermal Anomalies (LFTA) product is predicated on previous NASA-sponsored research to develop an active fire detection algorithm for the Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER) aboard Terra. ASTER multispectral imagery and derived active fire detection data have been routinely used over the past two decades to validate active fire products from MODIS and geostationary sensors. The LFTA product was added to FIRMS in August . Developed under the auspices of the NASA Applied Sciences Wildland Fire Program, the product is based on multispectral imagery acquired by the Operational Land Imager (OLI) sensor aboard the Landsat 8 and Landsat 9 satellites. The 30 m active fire detection data has 1,000 times and 160 times finer spatial resolution than the 1 km MODIS and 375 m VIIRS active fire detections, respectively. However, unlike MODIS and VIIRS, Landsat does not provide comprehensive global coverage on a daily basis. Each Landsat satellite typically images the same location on Earth every 16 days, but provides spatially explicit data and detail of observable fire activity at the time of the satellite overpass. These data augment fire management capabilities when available and can potentially enable applications at tactical scales.
The LFTA product is generated from all daytime OLI acquisitions and currently provided for most of North America (CONUS, southern Canada, and northern Mexico). Increased geographic coverage for the product is under consideration and will be investigated in the near future. Targeted nighttime OLI data collection for most of North America is also possible (OLI data are collected by request at nighttime), potentially increasing the opportunity for more frequent observations for selected geographic areas of heightened wildfire activity, but opportunities for nighttime observations are much more limited due to several factors.
What is the temporal frequency of Landsat observations and derived active fire data?
The OLI sensor aboard the Landsat 8 and Landsat 9 satellites each image the entire Earth every 16 days at the mid latitudes. The respective orbits of the two satellites are 8 days out of phase so when considered collectively, the entire planet is imaged at least every 8 days. Active fire detection data are derived from multispectral imagery acquired by OLI and provided daily for each of the scheduled WRS-2 paths imaged by Landsat 8 and 9 on a given date. Please consult the Landsat 8 and Landsat 9 orbit tracks on the FIRMS fire maps to visualize the time of Landsat observations for a selected date (as depicted below; please see How to View Satellite Orbit Tracks in the FIRMS FAQs). Additional active fire detection data from tasked nighttime Landsat OLI observations are possible, but not routinely conducted.
Image
The FIRMS fire map interface can display Landsat orbit tracks for user-selected dates with time of observation. True color and false color Landsat HLS imagery can also be displayed to view with active fire detection data. This display of a June 6, , Landsat 8 acquisition provides the user the coverage provided by a single 185 km Landsat swath.What is the spatial resolution of the Landsat observations and derived active fire data?
Each hotspot/active fire detection represents the center of a 30 m pixel flagged as containing one or more fires or other thermal anomalies (such as active volcanoes). The location is the center point of the pixel (not necessarily the coordinates of the actual fire).
What size of fires can be detected with Landsat?
Landsat provides better active fire detection performance compared to coarser spatial resolution sensors, such as MODIS and VIIRS. Flaming fires as small as approximately 4 m2 have a better than 50% chance of detection during the daytime observations. Nighttime detection of similar fire activity improves to an area as small as 1 m2.
Image
Schroeder, W., Oliva, P., Giglio, L., Quayle, B., Lorenz, E., & Morelli, F. (). Active fire detection using Landsat-8/OLI data. Remote Sensing of Environment, 185: 210-220. doi:10./j.rse..08.032Why are Landsat fire data only available for North America?
The Landsat Fire and Thermal Anomalies (LFTA) product is currently produced through a collaboration by USGS EROS Data Center, NASA, and the USDA Forest Service. Daily Landsat 8 and Landsat 9 OLI observations are processed in real-time by EROS and use the NASA-developed algorithm to produce active fire detection data that is ingested into FIRMS. Currently, only Landsat 8 and 9 OLI data collected by the USGS EROS Landsat Ground Station in Sioux Falls, South Dakota, are used to produce active fire detection data. The coverage of the USGS EROS Landsat Ground Station in Sioux Falls is highlighted in the figure below.
Expansion of the geographic coverage for the product is possible through the integration of additional near real-time Landsat 8 and Landsat 9 OLI data from other ground stations. The feasibility of including additional ground stations for North America and elsewhere is currently being investigated.
Image
Landsat ground stations for North America. While only the USGS EROS Landsat ground station in Sioux Falls, South Dakota currently produces LFTA data for the area covered by the USGS EROS Landsat Ground Station mask (within the red circle), both the USGS EROS Landsat ground station and the NOAA Gilmore Creek ground station (near Fairbanks, Alaska) collect Landsat 8 and Landsat 9 OLI data. The Canada Centre for Mapping and Earth Observation ground station (Canada) collects Landsat 8 OLI data. Recorded and real-time OLI data from Landsat 8 and Landsat 9 are collected by the EROS and Gilbert Creek ground stations when either satellite is within "line of sight". This condition is achieved when the Landsat orbit passes within the green circles on the map. Near real-time Landsat 8 OLI data can be collected by the Prince Albert ground station in Saskatchewan when Landsat 8 is within its "line of sight" (red circle).What is the temporal extent of Landsat Active Fire and Thermal Anomalies product in FIRMS?
The temporal extent of Landsat active fire data currently reaches back to June . Processing of earlier acquisitions by Landsat 8 and Landsat 9 is possible, but not planned at this time.
What are the attributes of the Landsat fire data?
The attributes are described in the following table:
Attribute Short Description Long Description Satellite Satellite L8 = Landsat 8; L9 = Landsat 9 Latitude Latitude (decimal degrees) Latitude of pixel center Longitude Longitude (decimal degrees) Longitude of pixel center Path Path Path number as identified in the World Reference System-2 (WRS-2) Row Row Row number as identified in the World Reference System-2 (WRS-2) Track Track The pixel location in the along-track dimension of the Landsat path based on the OLI Line-of-Sight (LOS) coordinate system. Generally increases in value from north to south for daytime and nighttime overpasses. Scan Scan The pixel location in the cross-track dimension of the Landsat path based on the OLI Line-of-Sight (LOS) coordinate system. Generally increases in value from west to east for daytime and nighttime overpasses. Acquire_Time Date and time of start of scan Date and time of acquisition/overpass of the satellite (in UTC). The format is YYYY-MM-DD followed by HH:MM-SS. For example, -07-23 10:09:00. Confidence Confidence Value domain: H, M, and L:What validation of the Landsat fire products has been performed?
Validation of the Landsat active fire product has primarily been performed using available high resolution aerial and commercial satellite imagery, coincident/near coincident 1 km Terra and Aqua MODIS active fire data, and 750 m and 375 m Suomi NPP VIIRS active fire data. A discussion of the validation procedures and results can be found in Schroeder, et al. ().
Where can I find more information about the Landsat fire data?
The Landsat active fire detection algorithm is based on the legacy algorithm developed for the Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER). Development of the Landsat algorithm was conducted under the auspices of the NASA Applied Sciences Program. Please refer to the following papers by Louis Giglio and Wilfrid Schroder and the NASA Applied Sciences Program project summary for more information on the Landsat active fire algorithm and product.
Schroeder, W., Oliva, P., Giglio, L., Quayle, B., Lorenz, E., & Morelli, F. (). Active fire detection using Landsat-8/OLI data. Remote Sensing of Environment, 185: 210-220. doi:10./j.rse..08.032
Giglio, L., Csiszar, I., Restás, Á., Morisette, J. T., Schroeder, W., Morton, D., & Justice, C. O. (). Active fire detection and characterization with the advanced spaceborne thermal emission and reflection radiometer (ASTER). Remote Sensing of Environment, 112(6): -. doi:10./j.rse..03.003
How does Landsat active fire data compare with MODIS and VIIRS active fire data?
Landsat active fire data are more than 1,000 times more resolute than MODIS and more than 150 times more resolute than VIIRS. The algorithm routinely detects fires as small as a few square meters or smaller, providing spatially explicit detail of the location and extent of fire activity at the time of a satellite overpass. Active fire fronts and scattered smaller fires within the fire perimeter are routinely detected. However, these relatively high spatial resolution data are provided at a lower temporal resolution compared to the daily observation frequency afforded by MODIS and VIIRS. Between both Landsat OLI instruments, areas in the lower to mid latitudes are imaged every 8 days. The temporal resolution of Landsat data can be increased though the collection of nighttime Landsat 8 and Landsat 9 imagery for targeted areas.
The following graphic shows Landsat active fire data compared for MODIS (Terra and Aqua) (1km) and VIIRS (Suomi NPP and NOAA-20) (375m) for the Double Creek Fire in Northeast Oregon on September 10, .
Image
Multi sensor imaging of the Double Creek Fire in Oregon on 10 September . Active fire detections, marked as yellow, derived from Landsat-9/OLI 30 m (displayed on a Landsat 9 HLS surface reflectance image), Terra/MODIS 1 km, Aqua/MODIS 1 km, Suomi NPP/VIIRS 375 m and NOAA-20/VIIRS 375 m (all displayed on respective MODIS and VIIRS corrected reflectance images). Temporal span for these consecutive images is approximately 3 hours.
Why does the latency differ between the Landsat active fire detection data and the HLS Landsat imagery?
Landsat active fire detection data are produced in collaboration with the USGS EROS Data Center in Sioux Falls, South Dakota. Daily Landsat 8 and Landsat 9 OLI imagery for the continental United States, southern Canada, and northern Mexico are captured via the Landsat ground station at EROS. This imagery is formatted and pre-processed in near real-time by the USGS and further processed onsite at EROS using NASA algorithms to extract 30-meter active fire detection data and made available for ingest into FIRMS approximately 30 minutes after a satellite overpass. Landsat imagery in FIRMS is processed as part of the Harmonized Landsat-Sentinel-2 (HLS) project to create BRDF-adjusted, 30-meter Landsat and Sentinel-2 surface reflectance products. Generated separately as provisional/research NASA products, HLS currently has a latency requirement of < 7 days. Consequently, FIRMS users desiring to view Landsat active fire detection data in the context of HLS Landsat true and false color composite imagery are required to wait until the processing of the HLS imagery is complete and made available in FIRMS. The HLS imagery is currently processed and available in FIRMS 2-4 days after acquisition, which is less than the current HLS processing latency requirement. Opportunities to reduce the latency of HLS imagery processing and its availability in FIRMS will continue to be assessed by NASA and implemented if possible.
Back to top
Can I download the fire data from the FIRMS Fire Map?
Active fire data are not currently available for download directly from the FIRMS Fire Maps. However, active fire locations from MODIS and VIIRS (and Landsat for the US and Canada) are available for download for the last 24 hours, last 48 hours, and last 7 days on the Active Fire Data page. Older MODIS and VIIRS data can be obtained through the Archive Download Tool.
You can create a customized view of imagery and geospatial data in the Fire Map viewer focused on a particular area of interest and share it with others as a PNG, JPG, GIF, or GeoTIFF image. To save and share an image of your map, click the screenshot icon at the bottom of the viewer and select your preferred option. Alternatively, you can create a unique URL for your customized view to share via or social media by clicking the share icon at the bottom of the Fire Map viewer.
What are the different sources of fire data in Fire Map?
See What are the different sources of data in FIRMS?
What does today vs 24 hours mean?
"Today" will show active fire detection data from GMT to present. 24 hours will display active fire detection data from GMT yesterday to present. Click on the "i" icon in the date tab (top right of the FIRMS Fire Map menu) to see what Today and 24 hours means in your local time.
What is the best way to view smoke in Fire Map?
To view smoke, add one of the corrected reflectance layers from the MODIS or VIIRS from the "backgrounds" section of the menu on the right. In addition, you could add the OMPS Aerosol Index (AI) and OMPS Aerosol Index (PyroCumuloNimbus) from the "overlays" section. OMPS AI detects and measures the density of smoke and suspended particles in the vertical air column in the atmosphere. Typical values of the aerosol index range from 0 to 5 and indicate dust storms or biomass burning smoke located in the lower troposphere (1-3 km). For AI values significantly higher than 5, use the OMPS Aerosol Index (PyroCumuloNimbus) layer. Larger AI values between 5 and 10 usually indicate dense smoke from intensely burning fires that reach higher in the troposphere. Once the AI gets above 10, the smoke has likely been produced from a pyroCb event, with dense smoke lofted into the upper troposphere and, often, into the stratosphere.
The OMPS AI layer is useful for identifying and tracking smoke from wildfires or biomass burning events as well as dust from desert dust storms and the long-range transport of volcanic ash from volcanic eruptions. Values of 5.0, or higher, indicate heavy concentrations of aerosols that could reduce visibility or impact human health.
Back to top
Can you notify me when a fire occurs in my area of interest?
We have developed a global fire alert system to notify users when a fire occurs in or near a specified area of interest, country, or protected area. You can subscribe to receive near-real time, daily, or weekly alerts in English, Spanish, or French.
To subscribe to, or learn more about the based alert system, please see the FIRMS Alerts page.
Do you provide mobile/cell text messages?
We do not currently provide SMS text messages. Previously, we helped develop such a service in collaboration with ESKOM and CSIR Meraka in South Africa for the protection of power lines in remote areas from wildfires by informing operators in the field about fire events in near-real time. For more information see: Davies, D. K., H. F. Vosloo, et al. (). Near real-time fire alert system in South Africa: from desktop to mobile service. Proceedings of the 7th ACM conference on Designing interactive systems Cape Town, South Africa ACM: 315-322. doi:10./.
What are the near-real time alerts?
The fire points are sent out in an as soon as they are processed by FIRMS. Alerts are sent only if a fire/hotspot was detected. The number of alerts varies according to geographic location: more frequent satellite observations occur at high latitudes. Near real-time alerts are usually sent within 3 hours of satellite observation for fires detected using MODIS aboard Aqua and Terra, and within 4 hours for fires detected using data from VIIRS aboard Suomi NPP and NOAA-20/21. Near real-time alerts are subscribed to, and managed by, the user just the same way as the daily and weekly detection summaries.
Go to the FIRMS Alerts page.
Go to the FIRMS Alerts page.
I requested a map image with my alert but I cant see the map in my alert, where is it?
Please check the settings in your client. Your client may be blocking images from being displayed in your as a security measure. You will have to enable the choice to view images in your . Your client may also be sending the alert to your spam/junk folder as a security measure; it is likely that you will also not be able to view the map image if the is in your spam/junk folder. You should add the FIRMS address and/or domain to the safe senders list so it will deliver the to your inbox and display images. You can also view the map by clicking on the link below the map image View Map Image on FIRMS server; this will open the map in your browser. Map images are stored for 14 days on the FIRMS server.
I requested a CSV file with my alert but I cant see the file attached to my alert, where is it?
If the number of fires in the alert exceeds 90,000, a link to download the CSV file will be provided instead of an attachment. CSV files are stored for 14 days on the FIRMS server.
I have received a CSV file as part of my Alert, can I add it as a layer in a Desktop GIS software?
A CSV (Comma Separated Value) fileis a text file in which separate fields are delimited by commas. This type of file can be used to store simple tabular data efficiently, minimizing file size. CSV files are easily opened with DB administration software such as PostgreSQL or MS Access, or by spreadsheet software such as MS Excel. This type of file can also be used to easily plot point data on desktop GIS software, given, as the active fire data does, that the tabular data contain X and Y coordinate information. The active fire data contains latitude and longitude location coordinates and the attributes of the detected fires.
Why was FIRMS US/Canada developed?
FIRMS US/Canada is the result of an agreement between the U.S. Forest Service and NASA to modernize the Forest Service's distribution of active fire information by leveraging FIRMS web-based active fire mapping tools and capabilities for disseminating data, products and services.
The significance of this NASA-Forest Service partnership goes beyond the creation of FIRMS US/Canada. It establishes a single, authoritative source of near real-time fire mapping, visualization, and geospatial data products and information for the United States and Canada jointly supported and enabled by both agencies. This enhanced cooperation will boost the availability of selected contextual geospatial data to increase the utility of near real-time fire geospatial data products, as well as increase opportunities, under the auspices of LANCE, to evaluate and integrate additional sources of NASA, NOAA, and international space agency satellite data of value to the user community.
What active fire data and imagery are available in FIRMS US/Canada?
*Other data and imagery sources are available for global coverage. See the FIRMS Global table.
1FDC: Fire Detection and Characterization algorithm
2The pixel size systematically grows from sub-satellite towards the edge of the disk
3FRPPIXEL: Fire Radiative Power algorithm
4Latency refers to the estimated time from satellite observation to availability in FIRMS
5Currently curated for FIRMS US/Canada
See Acronyms section for definitions.
What specific areas are covered by FIRMS US/Canada?
FIRMS US/Canada includes coverage for the continental United States, Alaska, Hawaii, Puerto Rico, and Canada. Due to the geographic extent of these areas of interest, Mexico and Caribbean countries are also included.
What are the differences between FIRMS Global and FIRMS US/Canada?
Just like FIRMS Global, FIRMS US/Canada provides active fire data generally within 3 hours of a satellite observation from MODIS aboard NASAs Aqua and Terra satellites and VIIRS aboard the Suomi NPP and NOAA-20 satellites. Imagery is typically available within 3 to 4 hours and can be viewed on the interactive Fire Map application.
In addition, FIRMS US/Canada meets the new Forest Service requirements by offering additional contextual layers and enhancements, including classifying fires to show time since detection to depict active fire fronts, incident locations and other information for current large fires in the US and Canada. FIRMS US/Canada provides current and historical corrected reflectance imagery from NASA and NOAA satellites, U.S. and Canada administrative ownership boundaries, daily fire danger forecasts, and current National Weather Service fire weather watch and red flag warning areas.
What is the difference between the USA Active Fires IRWIN and USA Active Fires IMSR layers displayed in FIRMS US/Canada, and what information do they provide?
The USA Active Fires layer includes wildland fire incidents from the Integrated Reporting of Wildland Fire Information (IRWIN) integration service that meet the following criteria: valid Wildfire (WF); Prescribed Fire (RX); and Incident Complex (CX) records that have not been declared contained, controlled, or out, have not had fire report records completed (certified), are not quarantined in IRWIN due to potential conflicts with other records, and have incident information has been updated within the past 14 days. In addition to the above criteria, FIRMS also limits the incidents displayed in the USA Active Fires layer to fires that are at least 100 acres(40 hectares)in size. Selected descriptive information, including fire name, latest reported size, and date/time of ignition are included in the attributes (available when a user clicks on a fire icon). The displayed wildfire location is typically based on the identified or assumed point of origin of the fire. The IRWIN service is maintained by the U.S. Department of Interior Office of Wildland Fire. IRWIN harvests and integrates information from numerous federal, state, and local wildfire reporting systems in near real-time to provide a single source of current information about all reported wildfire incidents.
The Incident Management Situation Report (IMSR) layer includes only large wildland fire incidents currently being tracked by the National Interagency Coordination Center (NICC). Large wildland fire incidents are classified by the NICC as 100 acres or larger in timber and slash fuel models, 300 acres or larger in grass or brush fuel models. Wildland fires that do not meet the large wildland fire criteria but have a Complex Incident Management Team (CIMT) assigned are also included in the IMSR. For Complex fires, the location included in the IMSR may not have any associated active fire detections because the location represents the location of the incident management center. The IMSR layer is updated weekly (Friday) by the NICC when the National Wildland Fire Preparedness Level is at Preparedness Level 1 and daily when at Preparedness Level 2 or higher. The IMSR typically represents a subset of the fires in IRWIN
What is the source of the Canada Active Fires layer and what information does it provide?
The Canada Active Fires layer is based on information from the Canadian Wildland Fire Information System Active Wildfires dataset coordinated and maintained by the Canadian Interagency Forest Fire Center (CIFFC) and Natural Resources Canada. Reported wildfire locations are updated and provided daily by partner provincial and territorial fire management agencies and Parks Canada. The Canada Active Fires layer displays the location of current wildfires greater than 100 acres (40 hectares) in size and selected descriptive information, including fire name/ID, latest reported size, and date/time of ignition.
Where can I get more information about current wildfires in the U.S. and Canada?
Additional information about current large wildfires in the United States can be found in the Incident Management Situation Report published by the National Interagency Coordination Center. Additional information about current large wildfires in Canada can be found in the National Fire Situation Report published by the Canadian Interagency Forest Fire Center (CIFFC).
What is the source for USA Fire Weather Watches and Red Flag Warnings and what do they mean?
The National Weather Service (NWS) is the source for fire weather watches and red flag warnings. A fire weather watch or red flag warning is issued by local/regional NWS forecast offices when the combination of dry fuels and weather conditions in designated NWS fire weather zones support extreme fire danger. Criteria for fire weather watches and red flag warnings vary by NWS office, but are typically based on the presence of fuel characteristics favorable for large fire growth, as determined by the land management agencies, and NWS-established critical thresholds for temperature, sustained winds, and relative humidity. Fire weather watches are issued up to 72 hours before these conditions are expected to occur while red flag warnings are issued when these conditions are expected to occur or are occurring within the next 24 hours. A similar layer for Canada is currently not available.
What is the source for the USA Forecasted Fire Danger layer and what information does it provide?
The Forecasted Fire Danger layer is based on the National Fire Danger Rating System (NFDRS), which produces several components and indexes (e.g., ignition probability, fire rate of spread, etc.) that collectively describe current and potential fire danger conditions. Daily forecasted fire danger updates are used by wildland fire management agencies to assess current fire danger at local and national levels and inform decisions on an areas fire protection needs. NFDRS fire danger is expressed as an adjective class rating that normalizes rating classes across different fuel models, indexes, and station locations.
What is the source for the Canada Forecasted Fire Danger layer and what information does it provide?
The Canada Fire Danger forecast layer is a relative index of how easy it is to ignite vegetation, how difficult a fire may be to control, and how much damage a fire may do. These ratings, updated and mapped daily, are based on the effects of fuel moisture and weather conditions on fire behavior as determined by the Canadian Forest Fire Weather Index (FWI) System. Fire danger conditions are classified by the provincial and territorial fire management agencies and the choice and interpretation of classes may vary between provinces.
What is the source for the United States wildfire perimeters and what does it display?
The US Fire Perimeter layer displays the extent of current wildfires based on perimeter data from the National Incident Feature Service (NIFS) maintained by the National Interagency Fire Center (NIFC). Incident management teams for individual wildfire incidents are the source for these data. Input data/information sources for routine perimeter updates conducted by incident management teams, as well as the update frequency, can vary.
How do I interpret/use wildfire perimeter data in the context of satellite active fire detection data?
Wildfire perimeters for individual fires are typically updated once daily using relatively large scale data (airborne infrared flights, airborne or terrestrial GPS, etc.), but may be updated more or less frequently in particular circumstances. MODIS and VIIRS active fire detection data introduce additional factors that can contribute spatial and temporal alignment between these data and perimeters for ongoing wildfire incidents. These factors include:
Are additional capabilities planned for FIRMS US/Canada?
Yes. The beta release of FIRMS US/Canada was completed in January . Ongoing development activities include introduction of more remote sensing layers in the fire map viewer, particularly additional sources of near real-time satellite imagery and fire-relevant science data products, and enhancements to data downloads and user-defined fire alerts.
More about the Ultra Real-Time (URT) and Real-Time (RT) data in FIRMS
Ultra Real-Time (URT) MODIS and VIIRS active fire detections are available for the US and Canada. These data are compiled from multiple direct readout ground stations in CONUS as part of a partnership with the Space Science and Engineering Center (SSEC) at the University of Wisconsin-Madison. MODIS and VIIRS micro granules collected via direct broadcast are processed in seconds and made available in FIRMS within 1-2 minutes of satellite observation. The attributes of these URT data have a "version" ending in URT. (To obtain the attributes, zoom in and click on a MODIS or VIIRS active fire detection in FIRMS Map viewer, and review the VERSION column).
Real-Time (RT) MODIS active fire detections are available through Direct Broadcast stations that process the data after the overpass has occurred. These data are generally available in FIRMS 20-25 minutes after satellite observation.
See the Questions about Real-Time (RT) and Ultra Real-Time (URT) active fire data section for more information.
Additional direct readout ground stations are being incrementally added to FIRMS. These additions will further increase the availability of URT MODIS and VIIRS data for North America collected from Terra, Aqua, Suomi NPP and NOAA-20.
Back to top
Back to top
If you want to learn more, please visit our website Granite Supply.