time_series_duration.Rmd
Two different shortening methodologies are proposed here. The first is to simply take the last n years in a given time series (e.g. 30, 20, 10) and compare the events detected in the overlapping period of time. This method will help to address the question of how the length of a time series may affect the creation of the seasonal signal and 90th percentile threshold (hereafter referred to as “the climatologies”), and therefore the events detected. This is an important consideration and one that needs to be investigated to ascertain how large the effect actually is.
The second technique proposed here is re-sampling. The issue with re-sampling the time series at the three different lengths is that it will prevent direct comparison of the resulting climatologies. Rather, re-sampling the time series in this way is useful for the comparison of events in the same time series detected with the differing climatologies. In order to better compare the re-sample climatologies, the following measurement metrics will be quantified:
After looking at these effects, the next stage of this investigation will be to look into best practices on how to consistently detect events when time series are not of optimal length.
Finally, this brings us to the direct consideration of the inherent decadal trend in the time series themselves. Ultimately, this tends to come out as the primary driver for much of the event detection changes over time (Oliver et al. 2018). To this end it would also be worthwhile to compare results for the different time series length with and without the long-term trends removed.
First prize for all of this research would be to develop an equation (model) that could look at a time series and determine for the user how best to calculate a climatology. It seems to me that an important ingredient must be the decadal (or annual) trend. So one would then need to take into account everything learned from the methodologies proposed above and investigate what relationship decadal trends have with whatever may be found.
In this section we will compare the detection of events when we simply nip off the earlier decades (i.e. 80’s and 90’s) in the three pre-packaged time series in heatwaveR
.
library(tidyverse)
library(broom)
library(heatwaveR)
library(lubridate) # This is intentionally activated after data.table
library(fasttime)
library(ggpubr)
library(boot)
library(FNN)
library(mgcv)
In order to control more tightly for the effect of shorter time series I am going to standardise the length of the 30 year time series as well. Because the built in time series are actually 33 years long I am going to nip off those last three years. The WMO recommends that climatologies be 30 years in length, starting on the first year of a decade (e.g. 1981). Unfortunately the OISST data from which the built-in time series have been drawn only start in 1982. For this reason we will set the 30 year climatology period as being from 1982 – 2011. To match the output of the shortened time series against this, 2011 will be taken as the last year for comparison and the data from 2012 onward will not be used.
# First put all of the data together and create a site column
sst_ALL <- rbind(sst_Med, sst_NW_Atl, sst_WA) %>%
mutate(site = rep(c("Med", "NW_Atl", "WA"), each = 12053))
# Then calculate the different clims
sst_ALL_clim <- sst_ALL %>%
nest(-site) %>%
mutate(clim_10 = map(data, ts2clm, climatologyPeriod = c("2002-01-01", "2011-12-31")),
clim_20 = map(data, ts2clm, climatologyPeriod = c("1992-01-01", "2011-12-31")),
clim_30 = map(data, ts2clm, climatologyPeriod = c("1982-01-01", "2011-12-31"))) %>%
select(-data) %>%
gather(key = decades, value = output, -site) %>%
unnest()
# Then calculate the different decadal trends
# See Schlegel and Smit 2016 for GLS model explanation/validation
decadal_trend <- function(df) {
if(df$decades[1] == "clim_10"){
df1 <- df %>%
filter(t >= as.Date("2002-01-01"),
t <= as.Date("2011-12-31"))
} else if(df$decades[1] == "clim_20"){
df1 <- df %>%
filter(t >= as.Date("1992-01-01"),
t <= as.Date("2011-12-31"))
} else if(df$decades[1] == "clim_30"){
df1 <- df %>%
filter(t >= as.Date("1982-01-01"),
t <= as.Date("2011-12-31"))
}
df2 <- df1 %>%
ungroup() %>%
mutate(date = floor_date(t, "month")) %>%
group_by(date) %>%
summarise(temp = mean(temp,na.rm = T)) %>%
mutate(year = year(date),
num = seq(1, n()))
dt <- round(as.numeric(coef(gls(
temp ~ num, correlation = corARMA(form = ~ 1 | year, p = 2),
method = "REML", data = df2, na.action = na.exclude))[2] * 120), 3)
return(dt)
}
sst_ALL_DT <- sst_ALL_clim %>%
mutate(decades2 = decades) %>%
group_by(site, decades2) %>%
nest() %>%
mutate(DT = map(data, decadal_trend)) %>%
select(-data) %>%
unnest() %>%
rename(decades = decades2)
# Quick peak at mean seas clims
# This may work better as a figure...
short_table <- sst_ALL_clim %>%
group_by(site, decades) %>%
summarise(seas_mean = mean(seas)) %>%
left_join(sst_ALL_DT, by = c("site", "decades")) %>%
separate(decades, into = c("char", "years")) %>%
select(-char) %>%
mutate(years = as.numeric(years)) %>%
arrange(site, -years)
knitr::kable(short_table, col.names = c("site", "years", "mean seas. temp.", "dec. trend"), align = "c",
caption = "Table showing the overall mean temperature for the seasonal climatologies detected at the three sites for the most recent 10, 20, or 30 years of data. Also shown are the decadal trends detected for the given 30, 20, or 10 year period. Note that the long-term heating trend in the Mediterranean (Med) appears to slow in the most recent decade. The other two sites show that the long-term heating trend only becomes more extreme the closer to present the data are.",
digits = 2)
site | years | mean seas. temp. | dec. trend |
---|---|---|---|
Med | 30 | 17.67 | 0.27 |
Med | 20 | 17.76 | 0.52 |
Med | 10 | 18.06 | 0.11 |
NW_Atl | 30 | 8.58 | 0.21 |
NW_Atl | 20 | 8.57 | 0.59 |
NW_Atl | 10 | 8.68 | 1.23 |
WA | 30 | 21.54 | 0.16 |
WA | 20 | 21.61 | 0.20 |
WA | 10 | 21.63 | 1.50 |
Quickly take note of the fact that for all three time series, the mean seasonal climatology becomes warmer the shorter (closer to the present) the period used is. This is to be expected due to the overall warming signal present throughout the worlds seas and oceans. For the Mediterranean only this warming trend appears to slow near the end of the time series. We’ll return to the impact of this phenomenon later.
# A clomp of functions used below
# Written out here for tidiness/convenience
# Calculate only events
detect_event_event <- function(df, y = temp){
ts_y <- eval(substitute(y), df)
df$temp <- ts_y
res <- detect_event(df)$event
return(res)
}
# Run an ANOVA on each metric of the combined event results and get the p-value
event_aov_p <- function(df){
aov_models <- df[ , -grep("decades", names(df))] %>%
map(~ aov(.x ~ df$decades)) %>%
map_dfr(~ broom::tidy(.), .id = 'metric') %>%
mutate(p.value = round(p.value, 5)) %>%
filter(term != "Residuals") %>%
select(metric, p.value)
return(aov_models)
}
# Run an ANOVA on each metric of the combined event results and get CI
event_aov_CI <- function(df){
label_levels <- levels(as.factor(df$decades))
# Run ANOVAs
aov_models <- df[ , -grep("decades", names(df))] %>%
map(~ aov(.x ~ df$decades)) %>%
map_dfr(~ confint_tidy(.), .id = 'metric') %>%
mutate(decades = rep(label_levels, nrow(.)/length(label_levels))) %>%
select(metric, decades, everything())
# Calculate population means
df_mean <- df %>%
group_by(decades) %>%
summarise_all(.funs = mean) %>%
gather(key = metric, value = conf.mid, -decades)
# Correct CI for first category
res <- aov_models %>%
left_join(df_mean, by = c("metric", "decades")) %>%
mutate(conf.low = if_else(decades == label_levels[1], conf.low - conf.mid, conf.low),
conf.high = if_else(decades == label_levels[1], conf.high - conf.mid, conf.high)) %>%
select(-conf.mid)
return(res)
}
# A particular summary output
event_output <- function(df){
res <- df %>%
group_by(decades) %>%
# select(-event_no) %>%
summarise_all(c("mean", "sd"))
return(res)
}
Given that there are perceptible differences in the mean seasonal trend values between the decades of data used, let’s see if an ANOVA determines these differences to be significant.
sst_ALL_clim_aov <- sst_ALL_clim %>%
select(-t, - temp, -doy) %>%
mutate(decades = as.factor(decades)) %>%
unique() %>%
group_by(site) %>%
nest() %>%
mutate(res = map(data, event_aov_p)) %>%
select(-data) %>%
unnest()
ggplot(sst_ALL_clim_aov, aes(x = site, y = metric)) +
geom_tile(aes(fill = p.value)) +
geom_text(aes(label = round(p.value, 2))) +
scale_fill_gradient2(midpoint = 0.1)
Seeing that there are no significant differences between the climatologies for the different time periods, let’s compare the results of the MHWs detected in the final decade of each time series using the different climatologies calculated using 1, 2, or 3 decades. We will make these comparisons using an ANOVA, as above.
# Calculate events and filter only those from 2002 -- 2011
sst_ALL_event <- sst_ALL_clim %>%
group_by(site, decades) %>%
nest() %>%
mutate(res = map(data, detect_event_event)) %>%
select(-data) %>%
unnest(res) %>%
filter(date_start >= "2002-01-01", date_end <= "2011-12-31") %>%
select(-c(index_start:index_end, date_start:date_end))
# ANOVA p
sst_ALL_aov_p <- sst_ALL_event %>%
select(site, decades, duration, intensity_mean, intensity_max, intensity_cumulative) %>%
nest(-site) %>%
mutate(res = map(data, event_aov_p)) %>%
select(-data) %>%
unnest() #%>%
# spread(key = metric, value = p.value)
# visualise
ggplot(sst_ALL_aov_p, aes(x = site, y = metric)) +
geom_tile(aes(fill = p.value)) +
geom_text(aes(label = round(p.value, 2))) +
scale_fill_gradient2(midpoint = 0.1) +
theme(axis.text.y = element_text(angle = 90, hjust = 0.5))
Whereas the results for the metrics do show some differences, they are not significant. The results do appear most different for the Mediterranean site and this is likely because, unlike the other two sites, the Mediterranean was not heating up as quickly in the most recent decade. This then means that when we compare only the events in the most recent decade of data, the Mediterranean results are most different because the other two have their larger events in the most recent decade. This is most evident in the Western Australia time series as we see p-values >=0.98 for three of the metrics. Specifically this is because of the monstrous 2010/11 MHW that occurred there.
# ANOVA CI
sst_ALL_aov_CI <- sst_ALL_event %>%
select(site, decades, duration, intensity_mean, intensity_max, intensity_cumulative) %>%
nest(-site) %>%
mutate(res = map(data, event_aov_CI)) %>%
select(-data) %>%
unnest()
CI_plot_1 <- ggplot(sst_ALL_aov_CI, aes(x = site)) +
geom_errorbar(position = position_dodge(0.9), width = 0.5,
aes(ymin = conf.low, ymax = conf.high, linetype = decades)) +
facet_wrap(~metric, scales = "free_y")
CI_plot_1
When we look at the confidence intervals (CI) we see that all of the MHW metrics for all of the time periods overlap rather well. As expected, we also see that the duration/intensity of MHWs tend to increase the more decades of data are used to create the climatologies.
Now knowing that the daily values that make up the climatologies do not differ, as seen with the ANOVA results, we also want to check if the distributions of the climatologies themselves differ. We will do this through a series of pair-wise two-sample KS tests.
Now that we know that the detected events do not differ significantly when different time-frames are used, we still want to know by what amounts they do differ. This information will then later be used during the re-sampling to see if this gap can be reduced through the clever use of statistics.
# Long data.frame for plotting
sst_ALL_event_long <- sst_ALL_event %>%
select(site, decades, duration, intensity_mean, intensity_max, intensity_cumulative) %>%
# select(-event_no) %>%
gather(variable, value, -site, -decades)
# Visualisations
# Notches are unhelpful and noisey here
ggplot(sst_ALL_event_long, aes(x = site, y = value)) +
geom_boxplot(aes(fill = decades), position = "dodge", notch = FALSE) +
facet_wrap(~variable, scales = "free_y")
The Mediterraean (Med
) data change over time in a more predictable, linear fashion. The Western Australia (WA
) data changed very little depending on the number of decades used for the climatology, which is most certainly due to that one huge event near the end. The Northwest Atlantic (NW_Atl
) results for the 10 and 20 year periods appear to generally be more similar than for the 30 year period.
With these benchmarks established, we will now move on to re-sampling to see if the effect of a shorter time series on the detected climatology can be mitigated.
With the effect of shortening time series on the detection of events quantified, we will now perform re-sampling to simulate one hundred 10, 20, and 30 year time series in order to quantify how much more variance one may expect from shorter time series. This will be measured through the following statistics:
The secondary goal of this step in this section of the methodology is to also identify how much more accurate this re-sampling may be able to make the climatologies generated from shorter time series as a technique for addressing this potential short-coming (yes, that was a pun).
### Need to insert here a function that does what 'sample_n' does
### But only selects from the appropriate range of dates
sst_repl <- function(sst) {
sst.sampled <- sst %>%
mutate(sample_10 = map(data, sample_n, 10, replace = TRUE),
sample_20 = map(data, sample_n, 20, replace = TRUE),
sample_30 = map(data, sample_n, 30, replace = TRUE))
return(sst.sampled)
}
parse_date_sst <- function(data, rep_col, len) {
parsed <- data %>%
mutate(id = rep_col,
y = year(t),
m = month(t),
d = day(t)) %>%
group_by(site, rep, doy) %>%
mutate(y = seq(2012-len, by = 1, len = len)) %>%
mutate(t = as.Date(fastPOSIXct(paste(y, m, d, sep = "-")))) %>%
select(-y, -m, -d) %>%
na.omit() # because of complications due to leap years
return(parsed)
}
sst_ALL_doy <- sst_ALL %>%
mutate(doy = yday(as.Date(t))) %>%
nest(-site, -doy)
sst_ALL_repl <- purrr::rerun(100, sst_repl(sst_ALL_doy)) %>%
map_df(as.data.frame, .id = "rep")
sample_10 <- sst_ALL_repl %>%
unnest(sample_10) %>%
parse_date_sst("sample_10", len = 10) %>%
group_by(site, id, rep) %>%
nest() %>%
mutate(smoothed = map(data, function(x) ts2clm(x, climatologyPeriod = c("2002-01-01", "2011-12-31")))) %>%
unnest(smoothed)
sample_20 <- sst_ALL_repl %>%
unnest(sample_20) %>%
parse_date_sst("sample_20", len = 20) %>%
group_by(site, id, rep) %>%
nest() %>%
mutate(smoothed = map(data, function(x) ts2clm(x, climatologyPeriod = c("1992-01-01", "2011-12-31")))) %>%
unnest(smoothed)
sample_30 <- sst_ALL_repl %>%
unnest(sample_30) %>%
parse_date_sst("sample_30", len = 30) %>%
group_by(site, id, rep) %>%
nest() %>%
mutate(smoothed = map(data, function(x) ts2clm(x, climatologyPeriod = c("1982-01-01", "2011-12-31")))) %>%
unnest(smoothed)
sst_ALL_smooth <- bind_rows(sample_10, sample_20, sample_30)
save(sst_ALL_smooth, file = "data/sst_ALL_smooth.Rdata")
# This file is not uploaded to GitHub as it is too large
# One must first run the above code locally to generate and save the file
load("data/sst_ALL_smooth.Rdata")
With 100 re-sampled climatologies and thresholds created, let’s take a moment to see if this much larger sample size differs significantly.
sst_ALL_smooth_aov <- sst_ALL_smooth %>%
rename(decades = id) %>%
select(-rep, -t, -temp, -doy) %>%
mutate(decades = as.factor(decades)) %>%
unique() %>%
group_by(site) %>%
nest() %>%
mutate(res = map(data, event_aov_p)) %>%
select(-data) %>%
unnest()
ggplot(sst_ALL_smooth_aov, aes(x = site, y = metric)) +
geom_tile(aes(fill = p.value)) +
geom_text(aes(label = round(p.value, 4))) +
scale_fill_gradient2(midpoint = 0.1)
# for each day-of-year (doy) in the climatology, calculate the SD of the climatological means of the 100 re-samplings;
sst_ALL_sd <- sst_ALL_smooth %>%
group_by(site, id, doy) %>%
summarise(seas_sd = sd(seas),
thresh_sd = sd(thresh),
var_sd = sd(var)) %>%
gather(variable, value, -site, -id, -doy)
ggplot(sst_ALL_sd, aes(x = doy, y = value)) +
geom_line(aes(colour = id)) +
facet_grid(variable ~ site)
Hmmm…. Interesting how the seasonal and threshold signals of increased summer variance comes through so nicely in the Med
data the fewer samples are taken. The other two time series produce somewhat strange signals. We can see that the 30 year, and to a lesser extent 20 year, re-sampled time series are much smoother than the 10 year re-sample. The WA
data are either very strange, or very exposed to particularly intense events.
# sst_ALL_sd %>%
# group_by(site, variable, id) %>%
# summarise(sd_mean = mean(value))
A quick glimpse at the overall mean of the SD values for each site and re-sample period shows how similar they are. Actually, there is not a linear relationship between increased re-sample size and decreased variance. That being said, an increased re-sample size does appear to produce a more stable variance profile. Meaning that even though the mean SD for the 10 year re-samples are very similar to the 20 and 30 year re-samples, to an extent this is because the variance is more variable, ultimately flattening itself out somewhat. This may be seen in the figure above how the slight peaks come out for the 10 year samples both above and below the other lines based on larger samples. Again, this appears almost negligible.
# Base 30 year clims
sst_ALL_clim_base <- sst_ALL_clim %>%
filter(decades == "clim_30") %>%
select(site:doy, seas:var, -decades) %>%
unique()
# For each doy, calculate the RMSE of the re-sampled means relative to the true climatology (i.e. the one produced from the 30-year long time series);
sst_ALL_error <- sst_ALL_smooth %>%
select(site:doy, seas:var) %>%
unique() %>%
left_join(sst_ALL_clim_base, by = c("site", "doy")) %>%
mutate(seas.er = seas.x - seas.y,
thresh.er = thresh.x - thresh.y,
var.er = var.x - var.y)
sst_ALL_RMSE <- sst_ALL_error %>%
group_by(site, id, doy) %>%
summarise(seas_RMSE = sqrt(mean(seas.er^2)),
thresh_RMSE = sqrt(mean(thresh.er^2)),
var_RMSE = sqrt(mean(var.er^2))) %>%
gather(variable, value, -site, -id, -doy)
ggplot(sst_ALL_RMSE, aes(x = doy, y = value)) +
geom_line(aes(colour = id)) +
facet_grid(variable ~ site)
# It appears as though the actual time series constructed from re-sampling,
# while useful for creating climatologies, are totally deurmekaar
# and do not actually lend themselves to the detection of heatwaves
# Therefore it is necessary to replace the 'temp' values in the smoothed data with the real values
sst_ALL_smooth_real <- sst_ALL_smooth %>%
select(-temp) %>%
left_join(sst_ALL, by = c("site", "t"))
# Then caluclate events using the many re-smapled clims on the real temperature data
sst_ALL_smooth_event <- sst_ALL_smooth_real %>%
group_by(site, id, rep) %>%
nest() %>%
mutate(res = map(data, detect_event_event)) %>%
select(-data) %>%
unnest(res) %>%
filter(date_start >= "2002-01-01", date_end <= "2011-12-31") %>%
select(-c(index_start:index_end, date_start:date_end))
save(sst_ALL_smooth_event, file = "data/sst_ALL_smooth_event.Rdata")
# This file is not uploaded to GitHub as it is too large
# One must first run the above code locally to generate and save the file
load("data/sst_ALL_smooth_event.Rdata")
# Rename for project-wide consistency
sst_ALL_smooth_event <- sst_ALL_smooth_event %>%
rename(decades = id)
# correspondence of detected events when using climatologies calculated from reduced time series vs. when using the full duration time series climatologies.
# ANOVA p
sst_ALL_smooth_aov_p <- sst_ALL_smooth_event %>%
select(site, decades, duration, intensity_mean, intensity_max, intensity_cumulative) %>%
nest(-site) %>%
mutate(res = map(data, event_aov_p)) %>%
select(-data) %>%
unnest() #%>%
# spread(key = metric, value = p.value) %>%
# select(names(select(sst_ALL_event, -decades, -event_no)))
# sst_ALL_smooth_aov_p
# visualise
ggplot(sst_ALL_smooth_aov_p, aes(x = site, y = metric)) +
geom_tile(aes(fill = p.value)) +
geom_text(aes(label = round(p.value, 2))) +
scale_fill_gradient2(midpoint = 0.1) +
theme(axis.text.y = element_text(angle = 90, hjust = 0.5))
# Long data.frame for plotting
sst_ALL_smooth_event_long <- sst_ALL_smooth_event %>%
select(-event_no) %>%
gather(variable, value, -site, -decades)
## NB: This takes waaaayy yo long to visualise...
# Visualisations
# Notches are unhelpful and noisey here
# ggplot(sst_ALL_smooth_event_long, aes(x = site, y = value)) +
# geom_boxplot(aes(fill = decades), position = "dodge", notch = FALSE) +
# facet_wrap(~variable, scales = "free_y")
##
# ANOVA CI
sst_ALL_smooth_aov_CI <- sst_ALL_smooth_event %>%
select(site, decades, duration, intensity_mean, intensity_max, intensity_cumulative) %>%
nest(-site) %>%
mutate(res = map(data, event_aov_CI)) %>%
select(-data) %>%
unnest()
# Plot
ggplot(sst_ALL_aov_CI, aes(x = site)) +
geom_errorbar(position = position_dodge(0.9), width = 0.5, colour = "black",
aes(ymin = conf.low, ymax = conf.high, linetype = decades)) +
geom_errorbar(data = sst_ALL_smooth_aov_CI, show.legend = F,
position = position_dodge(0.9), width = 0.5, colour = "red",
aes(ymin = conf.low, ymax = conf.high, linetype = decades)) +
facet_wrap(~metric, scales = "free_y")
Though I was not able to run the boxplot, because my computer kept falling over, the CI plot above demonstrates that the difference detected in the first round of experiments (when the real 10, 20, and 30 year baselines were used to generate a single clim each) holds up when we re-sample the clim generation 100 times. The centre around which the CI spread may be found remains very similar between the two experiments, with the distance between the upper and lower limits shrinking dramatically with re-sampling. Through re-sampling we see that there is a difference between the 10 year clims and the 20 and 30 year clims. But that there is no difference between the 20 and 30 year clims.
Note that in an earlier version of this vignette bootstrapping was also tested. It has since been removed as it was shown to not be effective. This was because the bootstrapping of random values to create climatologies created much lower values than the real data because while the bootstrapping does sample the data randomly, it then takes those n random samples and creates one mean value from them. This then makes artificially even values and so when one wants to calculate a 90th percentile threshold from this it is almost identical to the seasonal climatology.
In this section we want to look at how the categories in the different time periods compare. I’ll start out with doing basic calculations and comparisons of the categories of events with the different time periods. And then based on how that looks, see if I can think of a way of quantifying the differences.
# Calculate categories from events in one function for nesting
detect_event_category <- function(df, y = temp){
ts_y <- eval(substitute(y), df)
df$temp <- ts_y
res <- category(detect_event(df))
res$event_name <- as.character(res$event_name)
return(res)
}
First up we compare the category results for the events with the one reduced clim as done before.
# Calculate categories and filter only those from 2002 -- 2011
sst_ALL_clim_category <- sst_ALL_clim %>%
group_by(site, decades) %>%
nest() %>%
mutate(res = map(data, detect_event_category)) %>%
select(-data) %>%
unnest(res) %>%
filter(peak_date >= "2002-01-01", peak_date <= "2011-12-31") #%>%
# select(-c(index_start:index_end, date_start:date_end))
# Indeed... but how to compare qualitative data?
# Let's start out by counting the occurrence of the categories
sst_ALL_clim_category_count <- sst_ALL_clim_category %>%
group_by(site, decades, category) %>%
summarise(count = n()) %>%
mutate(category = factor(category,
levels = c("I Moderate", "II Strong",
"III Severe", "IV Extreme")))
ggplot(sst_ALL_clim_category_count, aes(x = decades, y = count, fill = decades)) +
geom_bar(stat = "identity", position = position_dodge()) +
geom_text(aes(label = count, y = count/2)) +
facet_grid(category~site)
Those are about the results I was expecting. Now let’s run this same code on our 100 re-samples we created earlier.
# Calculate categories and filter only those from 2002 -- 2011
sst_ALL_smooth_real_category <- sst_ALL_smooth_real %>%
group_by(site, id, rep) %>%
nest() %>%
mutate(res = map(data, detect_event_category)) %>%
select(-data) %>%
unnest(res) %>%
filter(peak_date >= "2002-01-01", peak_date <= "2011-12-31") #%>%
# select(-c(index_start:index_end, date_start:date_end))
save(sst_ALL_smooth_real_category, file = "data/sst_ALL_smooth_real_category.Rdata")
load("data/sst_ALL_smooth_real_category.Rdata")
# Indeed... but how to compare qualitative data?
# Let's start out by counting the occurrence of the categories
sst_ALL_smooth_real_category_count <- sst_ALL_smooth_real_category %>%
rename(decades = id) %>%
group_by(site, decades, category) %>%
summarise(count = n()/100) %>%
mutate(category = factor(category,
levels = c("I Moderate", "II Strong",
"III Severe", "IV Extreme")))
ggplot(sst_ALL_smooth_real_category_count, aes(x = decades, y = count, fill = decades)) +
geom_bar(stat = "identity", position = position_dodge()) +
geom_text(aes(label = round(count), y = count/2)) +
facet_grid(category~site)
Interestingly, when re-sampling for potential clims we actually see the detection of more larger category events with the 10 year periods than with either 20 or 30. As noted earlier, this re-sampling removes the decadal trend from the data, which then affects the creation of the climatologies being used. With a shorter period being used to create the threshold, it appears that this allows for the detection of ‘larger’ events. Remember that the real temperature values are being used to detect these events, not the re-sampled temperature values that were used to calculate the climatologies. This was because the re-sampled temperatures were too jumbled to detect events with reliably.
(AJS: I was thinking about the event categories. What we need to do is detect the top five events using the 30 yr climatology.
# Calculate all event categories, not just the last ten years as done above
sst_ALL_clim_category_ALL <- sst_ALL_clim %>%
group_by(site, decades) %>%
nest() %>%
mutate(res = map(data, detect_event_category)) %>%
select(-data) %>%
unnest(res)
# The top 5 events froom 30 years
sst_ALL_clim_category_30_top_5 <- sst_ALL_clim_category_ALL %>%
filter(decades == "clim_30",
peak_date >= as.Date("1982-01-01"),
peak_date <= as.Date("2011-12-31")) %>%
group_by(site) %>%
arrange(-i_max) %>%
slice(1:5)
# The top 5 events froom 20 years
sst_ALL_clim_category_20_top_5 <- sst_ALL_clim_category_ALL %>%
filter(decades == "clim_20",
peak_date >= as.Date("1992-01-01"),
peak_date <= as.Date("2011-12-31")) %>%
group_by(site) %>%
arrange(-i_max) %>%
slice(1:5)
# The top 5 events froom 10 years
sst_ALL_clim_category_10_top_5 <- sst_ALL_clim_category_ALL %>%
filter(decades == "clim_10",
peak_date >= as.Date("2002-01-01"),
peak_date <= as.Date("2011-12-31")) %>%
group_by(site) %>%
arrange(-i_max) %>%
slice(1:5)
Then, using the sections of time series that overlap with these events, use reduced time series, make climatologies, and see how those top five events compare in their event metrics to those detected using the shorter climatologies.
# RWS: This is a bit of a schlep as it requires a much more nuanced take on how to go about comparing relavent lengths of clim periods. The method used throughout here, that of the most recent 10, 20, or 30 years for the clim period won't work here and I hesitate to create a new methodology this late in the section...
# RWS: So for now I'm skipping forward to the next step as that will work with the current methodology.
Then we do the reverse. Detect events in the reduced time series, find the top five, and see what the matching ones in the full time series are like in comparison.
# Function for pulling out comparable events based on closest peak dates
peak_date_index <- function(df){
sst_30 <- sst_ALL_clim_category_ALL %>%
filter(decades == "clim_30")
peak_index <- as.data.frame(knnx.index(as.matrix(sst_30$peak_date),
as.matrix(df$peak_date), k = 1))
res <- sst_30[as.matrix(peak_index),]
}
# Find matching events based on nearest peak_date
sst_ALL_clim_category_10_top_5_match <- peak_date_index(sst_ALL_clim_category_10_top_5)
sst_ALL_clim_category_20_top_5_match <- peak_date_index(sst_ALL_clim_category_20_top_5)
sst_ALL_clim_category_top_5_match <- rbind(data.frame(sst_ALL_clim_category_10_top_5_match, match = "clim_10"),
data.frame(sst_ALL_clim_category_20_top_5_match, match = "clim_20"),
data.frame(sst_ALL_clim_category_30_top_5, match = "clim_30")) %>%
group_by(site, match, category) %>%
summarise(count = n()) %>%
mutate(category = factor(category,
levels = c("I Moderate", "II Strong",
"III Severe", "IV Extreme")))
ggplot(sst_ALL_clim_category_top_5_match, aes(x = match, y = count, fill = match)) +
geom_bar(stat = "identity", position = position_dodge()) +
geom_text(aes(label = count, y = count/2)) +
facet_grid(category~site)
With the differences in the count of categories for the detected events in the given time series lengths shown above, we will now compare the categories of the matching events between the different climatologies used.
sst_ALL_clim_category_10_top_5_compare <- rbind(data.frame(sst_ALL_clim_category_10_top_5, match = "clim_10"),
data.frame(sst_ALL_clim_category_10_top_5_match, match = "clim_30")) %>%
group_by(site, match, category) %>%
summarise(count = n()) %>%
mutate(category = factor(category,
levels = c("I Moderate", "II Strong",
"III Severe", "IV Extreme")))
ggplot(sst_ALL_clim_category_10_top_5_compare, aes(x = match, y = count, fill = match)) +
geom_bar(stat = "identity", position = position_dodge()) +
geom_text(aes(label = count, y = count/2)) +
facet_grid(category~site)
sst_ALL_clim_category_20_top_5_compare <- rbind(data.frame(sst_ALL_clim_category_20_top_5, match = "clim_20"),
data.frame(sst_ALL_clim_category_20_top_5_match, match = "clim_30")) %>%
group_by(site, match, category) %>%
summarise(count = n()) %>%
mutate(category = factor(category,
levels = c("I Moderate", "II Strong",
"III Severe", "IV Extreme")))
ggplot(sst_ALL_clim_category_20_top_5_compare, aes(x = match, y = count, fill = match)) +
geom_bar(stat = "identity", position = position_dodge()) +
geom_text(aes(label = count, y = count/2)) +
facet_grid(category~site)
So, compare specific events in addition to the way in which you have done it. Then do the same for median-sized events,
# RWS: I'm putting this on hold for now until I've editted the text here and in the skeleton.
and the smallest ones.
# RWS: Same for this chunk.
Maybe even those at the 2nd and 5th quantiles.)
# RWS: Same for this chunk.
With the amount of variance that may be accounted for through re-sampling and bootstrapping known, we will now look into how we may go about more confidently creating a climatology that will consistently detect events as similarly as possible by experimenting with how the various arguments within the detection pipeline may affect our results, given the different lengths of time series employed. After this has been done we will look into using the Fourier transform climatology generating method (https://robwschlegel.github.io/MHWdetection/articles/Climatologies_and_baselines.html) to see if that can’t be more effective. The efficacy of these techniques will be judged through a number of statistical measurements of variance and similarity.
Seeing as how 20 and 30 year periods produce very similar results, we will be focussing primarily on what may be done about the 10 year period to make it more similar to the 30 year period. I don’t think it is necessary to do so for the 20 year period.
Lastly we will now go about reproducing all of the checks made above, but based on a climatology derived from a Fourier transformation, and not the default methodology.
Oliver, Eric C.J., Markus G. Donat, Michael T. Burrows, Pippa J. Moore, Dan A. Smale, Lisa V. Alexander, Jessica A. Benthuysen, et al. 2018. “Longer and more frequent marine heatwaves over the past century.” Nature Communications 9 (1). doi:10.1038/s41467-018-03732-9.