Processing math: 100%

Surface Weather Parameters Forecasting Using Analog Ensemble Method over the Main Airports of Morocco

利用相似集合方法预报地面天气参量:在摩洛哥机场的应用

+ Author Affiliations + Find other works by these authors

PDF

  • Surface weather parameters detain high socioeconomic impact and strategic insights for all users, in all domains (aviation, marine traffic, agriculture, etc.). However, those parameters were mainly predicted by using deterministic numerical weather prediction (NWP) models that include a wealth of uncertainties. The purpose of this study is to contribute in improving low-cost computationally ensemble forecasting of those parameters using analog ensemble method (AnEn) and comparing it to the operational mesoscale deterministic model (AROME) all over the main airports of Morocco using 5-yr period (2016–2020) of hourly datasets. An analog for a given station and forecast lead time is a past prediction, from the same model that has similar values for selected predictors of the current model forecast. Best analogs verifying observations form AnEn ensemble members. To picture seasonal dependency, two configurations were set; a basic configuration where analogs may come from any past date and a restricted configuration where analogs should belong to a day window around the target forecast. Furthermore, a new predictors weighting strategy is developed by using machine learning techniques (linear regression, random forest, and XGBoost). This approach is expected to accomplish both the selection of relevant predictors as well as finding their optimal weights, and hence preserve physical meaning and correlations of the used weather variables. Results analysis shows that the developed AnEn system exhibits a good statistical consistency and it significantly improves the deterministic forecast performance temporally and spatially by up to 50% for Bias (mean error) and 30% for RMSE (root-mean-square error) at most of the airports. This improvement varies as a function of lead times and seasons compared to the AROME model and to the basic AnEn configuration. The results show also that AnEn performance is geographically dependent where a slight worsening is found for some airports.
    地面天气参量对所有领域(航空、海洋交通、农业等)的所有用户具有高度的社会经济影响和战略意义。然而,这些参量主要是通过确定性数值天气预报(NWP)模式进行预测的,其中包含大量的不确定性。本研究的目的是,利用相似集合方法(AnEn)以低成本集合计算改进这些参量的预报,并利用5年期间(2016–2020年)的每小时观测数据集,将其与摩洛哥主要机场的业务预报中尺度确定性模型(AROME)结果进行比较。对于一个给定的气象站和预报时效,某相似预报来自于过去的预报,基于相同的模型,对当前预报模式选定的预测因子具有相似的值。经观测验证的最好的相似预报构成了AnEn的成员。为描绘季节相关性,确立了两种配置;一种是基本配置,其中相似预报可能来自过去的任何日期;另一种是限制配置,其中相似预报位于目标预报周围某天的窗口期。此外,利用机器学习技术(线性回归、随机森林和XGBoost)开发了一种新的预测因子权重策略。这种方法有望完成相关预测因子的选择并找到它们的最佳权重,从而保证所使用的天气变量的物理意义和相关性。结果分析表明,开发的AnEn系统显示出良好的统计一致性,在摩洛哥的大多数机场,它在时间和空间上显著提高了确定性预测性能,Bias(平均误差)和RMSE(均方根误差)分别提高50%和30%。与AROME模型和AnEn基本配置相比,这种改进随着预测时间和季节的变化而变化。结果还表明,AnEn的性能与地理位置有关,在某些机场稍有恶化。
  • In most national meteorological services, numerical weather prediction (NWP) is mainly based on determinism. In this doctrine, given an initial state of the atmosphere, its evolution numerically leads to a unique prediction scenario. However, deterministic weather forecasts show several uncertainties, occasioned by several error sources related to model formulation (Orrell et al., 2001), initial state (PaiMazumder and Mölders, 2009), physical parameterization (Palmer, 2001), and lateral boundary conditions (Eckel and Mass, 2005). In front of those limitations, ensemble prediction takes advantage of all these uncertainty sources to construct multiple forecasts starting from slightly different but equally-probable initial states (Leith, 1974). Several national weather services around the world use ensemble prediction systems such as the NCEP (Toth and Kalnay, 1993, 1997; Toth, 2001; Zhu, 2005; Zhou et al., 2017), ECMWF (Buizza, 2008; Buizza and Richardson, 2017), the Meteorological Service of Canada (Buizza et al., 2005), and Météo-France (Vié et al., 2011). Those systems exhibit a high efficiency helping forecasters of those centers in the decision-making process. In order to construct ensemble prediction systems, several techniques were deployed such as perturbative methods that depend on atmospheric flow and based on perturbations in sub-spaces where initial condition errors grow faster, for example: breeding vectors (Toth and Kalnay, 1993, 1997) and singular vectors (Molteni et al., 1996). Recently, new data assimilation methods emerged and are included in ensemble systems for NCEP (Whitaker et al., 2008; Zhou et al., 2017) and for ECMWF (Hamill et al., 2000, 2011; Buizza et al., 2008).

    Among those techniques, analog ensemble (hereafter AnEn) forecasting is considered as an intuitive and low cost method of generating ensemble members. The fundamental idea of this method is to construct an ensemble forecast from a set of past observations of the variable to be predicted, neatly selected from a historical training dataset. For a given location, the most similar past forecasts to the current prediction are identified and their associated past observations are nominated as members of the analog forecast ensemble. Thus, the availability of records of NWP deterministic forecasts is a cornerstone to the application of this method. Analog ensemble method’s objective is to discern all the past weather conditions where the error probability density function was similar. Once those conditions are distinguished, one can use the past observation errors to deduce the future errors probability density function. Analog ensemble method was theoretically introduced firstly by Hamill and Whitaker (2006), and then, it was successfully applied by Delle Monache et al. (2013), hereafter DM13, to generate probabilistic prediction of wind at 10 m and temperature at 2 m. Then, several successful applications of this technique chained mainly in renewable energy for wind and solar energy (Alessandrini et al., 2015; Davò et al., 2016), energy load (Alessandrini et al., 2015), tropical cyclones intensity (Alessandrini et al., 2018), air quality predictions (Djalalova et al., 2015; Delle Monache et al., 2018), dynamical forecast errors correction (Yu et al., 2014; Gong et al., 2016), and also in the field of data assimilation (Lguensat et al., 2017). However, the previous studies focused on a few surface weather parameters. These parameters represent an important aspect of meteorology because they are the basis for all weather safety messages, weather forecasts, and weather warnings worldwide. Thus, extending research studies to assess the potential of AnEn method for more surface weather forecasting (e.g., 2-m relative humidity and mean sea level pressure, zonal and meridional 10-m wind components) is still needed.

    It is evident that the likelihood of finding good analogs depends strongly on the similarity metric and also the neighboring criteria (time window) and the weighting strategy applied to the predictors that exhibit correlations to the predictand. Several weighting algorithms of different philosophies were explored. The ultimate goal of these algorithms is to attribute a weight to each predictor in the AnEn method (Delle Monache et al., 2013). Junk et al. (2015) developed a static and dynamic weighting strategy where predictors weights are defined by probabilistic score minimization process over all plausible weights combinations, given discrete weight values. These strategies revealed an improvement of up to 20% in the performance of AnEn compared to the basic algorithm of DM13 where weights were initiated to 1 (all predictors have the same weight). On the other hand, Gensler et al. (2016) introduced a weight optimization wrapper technique based on the Nelder–Mead simplex algorithm (Lagarias et al., 1998). The main idea of this approach is to find the best weights combination selected randomly, which minimizes the root-mean-square error (RMSE) between AnEn forecasts and actual measurements. This method outperforms the DM13 strategy by 10% to 21%. Finally, Tuba and Bottyán (2018) explored another family of weighting methods, based on AHP (Analytic Hierarchy Process) decision-making algorithm (Saaty, 1987). This method reproduces the best weights combination based on eigenvector decomposition of a matrix of ratios. Those ratios translate the importance of each variable relatively to others. Weights in this method are represented by the normalized eigenvector associated with the maximal eigenvalue. A novelty in our study is the usage of machine learning techniques to compute the predictors weights. This strategy is investigated and evaluated against traditional statistical method (linear regression). In addition, in this study we extend the target predictands and predictors to eight surface weather parameters: temperature and relative humidity at 2 m, surface and mean sea level pressure, wind speed and direction at 10 m, and also the zonal and meridional 10-m wind components.

    It is to mention that this work is the first AnEn application using mesoscale operational model AROME (Applications de la Recherche à l’Opérationnel à Méso-Echelle) (Seity et al., 2011). The latter is widely used among an important part of the scientific community, and it is also used operationally in many countries, including Morocco (Hdidou et al., 2020). In the literature, AROME model was subject of many scientific contributions especially in targeting atmospheric convective activity and data assimilation (Degrauwe et al., 2016; Sahlaoui et al., 2020). However, some scientific studies addressed the AROME-based ensemble forecasting using other approaches than AnEn (Vié et al., 2011; Bouttier et al., 2016; Bousquet et al., 2020). Thus, our research aims to apply AnEn to AROME outputs, particularly regarding the main surface weather parameters. In addition, the current operational AROME model was coupled with the global NWP ARPEGE (Action de Recherche Petite Echelle Grande Echelle) model (Courtier et al., 1991) since 2020 instead of the ALADIN (Aire Limitée Adaptation Dynamique Développement International) model (Bubnová et al., 1995) since 2015. The impact of lateral boundary conditions change on AnEn performances will be also assessed in this study since AnEn leverages a single deterministic NWP model to produce ensemble forecasting for surface weather parameters.

    In this study, we propose an in-depth analysis of different configurations applied to AROME-based analog ensemble forecasting. To test the performance of the developed ensemble forecasting system and to determine the optimal configuration, several experiments using different scenarios are performed. The experiments use more meteorological variables as predictors and provide a 25-member analog ensemble forecast. The results analysis is investigated using 15 meteorological stations, located mainly in airports, in Morocco over 1-yr period. The deterministic AROME forecasts are used as a reference in order to better understand the analog method impact on the surface weather forecasts.

    This paper is organized as follows. Section 2 describes the study domain and datasets with a brief overview on the used NWP model AROME. Section 3 is dedicated to the methodology and the experimental design. Results analysis based on verification scores of the global performance is detailed in Section 4 regarding the spatiotemporal distribution and also the season dependency. This paper ends with a discussion section and conclusions, which summarize the main findings of this research.

    The mesoscale limited area model used in this study is the AROME-Morocco model (Hdidou et al., 2020), which has been in operational use at the Moroccan National Meteorological Service since 2015. This model was developed by Météo-France (Seity et al., 2011) and is being maintained and further refined in collaboration between the meteorological institutes belonging to the ALADIN-HIRLAM consortia (http://www.umr-cnrm.fr/accord/).

    The AROME canonical model configuration has been developed to run in the convection-permitting resolutions starting from 2.5-km resolution. Its setup is described by Seity et al. (2011) and Brousseau et al. (2016). Its physical parameterizations come mostly from the Méso-NH research model (Lafore et al., 1998) whereas the dynamical core is the non-hydrostatic ALADIN core (Bubnová et al., 1995). The coupling between the atmosphere and the underlying surface was based on the SURFEX system (www.umr-cnrm.fr/surfex/).

    AROME-Morocco covers the Moroccan kingdom and the south of Spain with a horizontal resolution of 2.5 km and 90 vertical levels, with the lowest level at about 5 m above the ground. The lateral boundary conditions are provided by hourly forecasts from the ALADIN model (Bubnová et al., 1995) and from the ARPEGE model (Courtier et al., 1991) since January 2020. AROME-Morocco runs operationally twice a day with a 48-h forecast range.

    The dataset used in this work comprises eight meteorological parameters: temperature (T2m) and relative humidity (RH2m) at 2 m, mean sea level pressure (MSLP), surface pressure (SURFP), wind speed (WS10m) and direction (WD10m) at 10 m, zonal (ZW10m) and meridional (MW10m) components of 10-m wind. The study period covers 5 yr (2016–2020). The hourly forecasts of these parameters are extracted from midnight AROME run outputs up to 24 h of every day. Similarly, the same parameters are made available hourly from SYNOP (surface synoptic observations) messages of 15 national airports covering the Moroccan territory (see Fig. 1) and are spatially distributed over topographically heterogeneous terrain. It should be noted here that airports in mountainous regions are excluded from this study because the geopotential parameter is archived instead of MSLP.

    Fig  1.  Map showing orography in meters and the position of the synoptic meteorological stations used in this study, mainly located in airports, and being operational along the whole day (24 h).

    Morocco is a country in the subtropical zone of North West Africa. It is characterized by very different climates depending upon the subregion. The Moroccan climate is influenced by the Atlantic Ocean to the west, the Mediterranean Sea to the north, the dry Saharan air to the south and is locally modulated by the orographic effects induced by the Atlas Mountains (see Fig. 1). These factors have a strong impact on the variability of moisture and other surface weather parameters (Knippertz et al., 2003).

    As shown in Fig. 2, the basic idea behind analog ensemble method is to find synonymous weather situations to the current one. To achieve this goal, we use the past forecasts database provided by a deterministic NWP model, which form analogs dataset (step 1 in Fig. 2), and a time series of analogs verifying observations, which will be ensemble members (step 2 in Fig. 2), over a given location. Taken all together, these observations constitute the ensemble prediction for the current forecast (step 3 in Fig. 2). Then, the deterministic prediction value of the predictand is the mean of the ensemble. A weather situation may extend from a few hours to a few days (weather regimes predominate the weather situation for several days). Ordinarily the forecast period of analog method is very short, often up to 6 h (Horton et al., 2017), 12 h (Riordan and Hansen, 2002), and rarely 24 h (Hansen, 2007), since analog methods often work correctly on homogeneous weather conditions which is elementary for this type of forecast. The dataset of past forecasts imperatively contains a set of meteorological variables used as predictors.

    Fig  2.  Analog ensemble algorithm steps with day neighborhood and search restrictions.

    For a given location and time, one seeks to predict a variable from the set named as predictand, and uses all the eight available variables as predictors including the predictand. For example, naming the temperature at 2 m as predictand, the predictors are: T2m, RH2m, MSLP, SURFP, WS10m, WD10m, ZW10m, and MW10m from the past forecast dataset.

    In order to select the potential analogs, the basic similarity metric of Delle Monache et al. (2013) has been modified and used in this study. The basic DM13 metric is defined as follows:

    ||Ft,At||=Nνi=1wiσi˜tj=˜t(Fi,t+jAi,t+j)2, (1)

    where t is the current NWP deterministic forecast valid at the future time t at a station location; At is an analog (past forecast) at the same location and with the same forecast lead time but valid at a past time t′; Nν and wi are the number of physical variables used in the analogs search and their weights, respectively; σi is the standard deviation of the time series of past forecasts of a given variable i at the same location and forecast lead time; ˜t is equal to half the number of additional times over which the metric is computed; and Fi,t+j and Ai,t+j are the values of the forecast and the analog in a time window with j from −˜t to +˜t for a given variable i.

    To assess the impact of the day neighboring on the AnEn performance, this criteria is added to DM13 formula in order to capture flow dependency and to force analogs to be in the same season and at a near date to the future forecast (target). Thus, Eq. (2) used in the modified version of DM13 is defined as follows:

    ||Ft,At,k||=Nνi=1wiσi˜tj=˜t(Fi,t+jAi,k,t+j)2, (2)

    where subscripts t and t′ represent the lead times of a forecast in the future (F) and in the past (A, i.e., a potential analog) respectively. The subscript ˜t is equal to half temporal window of additional times over which the distance is calculated; and Fi,t+j and Ai,k,t+j are the values of the forecast and the analog in a time window of 2 × ˜t h and 2 × ˜d days for a given variable i, where ˜d is the day window for the daily neighborhood configuration and k [target_date − ˜d, target_date + ˜d].

    Analog search dataset (training) is restricted only to days within the day neighborhood in previous years. To retrieve the basic formula of DM13 similarity metric where AnEn looks for analogs throughout the available historical dataset, it suffices to drop daily neighborhood index k from Eq. (2).

    In contrast with the DM13 method, which uses few predictors and where predictors weights are all equal and initiated to 1, hereinstead, for every predictand from the observed features, the set of predictors is constructed from all available eight parameters. In addition, a machine learning based approach is used as a weighting strategy. Thus, three machine learning models are constructed (Table 1), tuned, and cross validated all over the training period to find the optimal weights used as input to AnEn.

    Table  1.  Brief description of the machine learning techniques used for computing predictors weights
    MethodDefinition
    Linear regressionIn machine learning literature, regression points out to determine linear relations between the independent and dependent variables recursively via minimization of errors in predictions made by the linear model (Hope, 2020)
    Decision treesDecision trees are a powerful non-parametric supervised machine learning algorithm for classification and regression. The main idea of the algorithm is to best split the dataset into smaller and smaller subsets to predict the target value. The splitting process at each node continues until no further gain can be made or a preset rule is reached. Quality of splits is measured by impurity criteria; for regression one uses variance or mean square error (MSE) reduction to decide to split a node in two or more subnodes. Suppose we are developing a binary tree, the algorithm will first pick a value, and split the data into two subsets. For each subset, it will calculate the MSE separately. The tree chooses the value which results in the smallest MSE value
    Decision trees based techniques (random forest and XGBoost)These techniques merge several decision tree models named estimators to produce an optimum predictive model. This empowers better performances in contrast to a single model. Those methods can be split into two main algorithm families:
    • Boosting and bagging techniques like XGBoost (Chen and Guestrin, 2016) that uses gradient descent minimization on decision trees to create multiple models combined recursively while correcting previous ones to generate a final optimal model. XGBoost is very efficient in terms of usage of computational resources and processing speed (Jhaveri et al., 2019).
    • Random techniques like random forest that stems from decision trees and uses randomness to optimize accuracy by creating a forest of decision trees and merges them to form a more accurate model (Breiman, 2001; Jhaveri et al., 2019)
     | Show Table
    DownLoad: CSV

    The best model found for each technique is used to find a generic equation that links the predictand to predictors (called also features), and then predictors importance’s coefficients are inferred. The importance coefficients are then scaled and normalized to form predictors weights in AnEn. Neural networks technique was excluded from the benchmark due to the hardness in physical interpretation of importance coefficients in output. Features importance is variously calculated from a machine learning techniques family to another. The minimization method is a centerpiece in the process. The ordinary least squares method is usually used to find the weights that minimize the squared differences between the actual and the estimated outcomes as in Eq. (3).

    ˆα=argminα0,,αpni=1[y(i)(α0+pj=1αjx(i)j)]2, (3)

    where y(i) is the ith observation for the predictand y, x(i)j refers to the ith value for predictor xj, and αj is the linear coefficient of predictor xj. For linear regression models predictors importance can be measured by the absolute value of its t-statistic. The t-statistic tˆαj is the estimated weight ˆαj scaled with its standard error SE(ˆαj) as shown in Eq. (4).

    tˆαj=ˆαjSE(ˆαj). (4)

    For the decision tree based methods, feature importance is defined as the decrease in node impurity times the probability of reaching that node. The node probability equals the number of samples that reach the node, divided by the total number of samples. The higher the value the more important the variable. At each split in each tree the improvement in the split criterion—here is mean square error—is the importance measure attributed to the splitting variable and is accumulated over all the trees in the trees ensemble (forest) separately for each variable.

    For each meteorological parameter as a target predictand (T2m, RH2m, MSLP, SURFP, WS10m, WD10m, ZW10m, and MW10m), two configurations of daily neighboring are the basic DM13 version where analogs are selected from any date regardless of their season or month; and the modified version proposed in this study, where we select analogs imperatively from a day window around the day of the target. By this manner, we make sure that analogs belong to the same season as the target. This aims to assess the flow dependency of the AnEn performance, thus gaining time for operational use when looking for the best analogs.

    For each configuration, a weight optimization of the eight AnEn predictors was performed over all studied airports using the three techniques (XGBoost, random forest, and linear regression), for the 1–24 forecast hours over the training period (2016–2018). For all airports, the combination leading to the lowest mean square error of the machine learning (ML)-developed model is chosen and kept constant over the testing period (2019). The set of eight predictors were used because they are considered relevant for the prediction of the target variable and are the main available parameters in SYNOP messages. This gives a potential for this to be used worldwide. Additional variables, which have not been tested in this study, could further improve the results. The DM13 basic weighting strategy is also performed and used as a benchmark.

    As a result, 64 configurations (8 parameters × 2 neighborhood configurations × 4 weight optimization strategies) have been performed over all the 15 airports using the hourly AROME forecasts at the nearest grid point and the equivalent observations (Table 2). Each configuration provides a 25-member analog ensemble forecast. All the possible combinations are defined with the constraint iwi=1, where wi is the predictors weights. For the second category of experiments, the same configurations have been carried out, over the training period 2016–2019, to assess the impact of the boundary condition change on the AnEn performance over the testing period (2020).

    Table  2.  Experimental setup with more details about the used configurations
    ParameterExperiment 1Experiment 2
    Training period2016–20182016–2019
    Verification period20192020
    Boundary conditions changeNoYes
    Location15 synoptic airports of Morocco
    NWP modelAROME 41 cycle, 2.5 km, 90 VLs.
    Ensemble prediction system (EPS) techniqueAnalog ensemble
    Dataset24 hourly observations and forecasts
    FeatureT2m, RH2m, WS10m, WD10m, ZW10m, MW10m, SURFP, and MSLP
    Weighting strategyNo weight/random forest/linear regression/XGBoost
    Time window±1 h
    Day window±15 days
    ConfigurationDM13/DM13 with restriction
     | Show Table
    DownLoad: CSV

    In this section, we analyze first the weight optimization results. Then, the AnEn’s performance with the different configurations described previously is assessed and compared to the original version of DM13 using common verification metrics for evaluation of deterministic predictions, namely Bias (mean error) and RMSE. These metrics are described with more details in Jolliffe and Stephenson (2003) and Wilks (2005). However, as an ensemble forecast, AnEn prediction system must be statistically consistent with the observations in the large scale flow given the study domain is large. This statistical consistency is assessed in this study using the spread-error diagram.

    Many research studies on AnEn assign equal weights wi=1 to each predictor (Delle Monache et al., 2013; Delle Monache, 2015). To take into account the strength of the relationship between individual predictors and the target variable, the predictors weight optimization has been performed based on three machine learning techniques (XGBoost, random forest, and linear regression).

    Table 3 shows the percentage of occurrence of all possible weight values over all airports for each predictor, where T2m is taken as predictand. It is seen clearly from this table that the past forecasted T2m is the more relevant predictor, as expected, followed by MSLP and SURFP with its weights taking values greater than 0 at 48.4% of the nearest grid points from the studied airports.

    Table  3.  Percentage of occurrence of possible weights values for each predictor for T2m as predictand
    VariableWeighting value
    0 0.10.20.30.4 0.5 0.6 0.7 0.80.91
    T2m (%)19.4 9.76.5 3.212.922.625.8
    RH2m (%)58.141.9
    WS10m (%)83.916.1
    MSLP (%)51.619.43.29.716.1
    SURFP (%)61.3 6.56.59.716.1
    WD10m (%)83.916.1
    MW10m (%)74.222.63.2
    ZW10m (%)7129
     | Show Table
    DownLoad: CSV

    Results analysis for the other variables shows the relevance of MSLP and SURFP as predictors for all of them indicating the influence of large scale atmospheric conditions on each predictor. In addition, it is found that the zonal wind component also takes an important weight for RH2m as predictand. In fact, RH2m is influenced by the zonal circulation of air-masses and atmospheric flows. For zonal and meridional 10-m wind components, it is also found that 10-m wind speed is a relevant predictor. This is obvious by the nature of construction of this predictor, which is equal to the root square of the sum of wind components squares. While each variable is highly correlated with itself as predictor, some discrepancies are found between the three machine learning techniques regarding the order of high weights affectation to predictors.

    Applying this static weighting strategy makes it possible to select the potential AnEn predictors for probabilistic target variable forecasting and also to find their optimal weights. This could improve the AnEn performance since their selection is based on physical link between predictors and target variables.

    An especially important aspect of ensemble forecasting is its capacity to yield information about the magnitude and nature of the uncertainty in a forecast. While dispersion describes the climatological range of an ensemble forecast system relative to the range of the observed outcomes, ensemble spread is used to describe the range of outcomes in a specific forecast from an ensemble system. Qualitatively, we have more confidence that the ensemble mean is close to the eventual state of the atmosphere if the dispersion of the ensemble is small. Conversely, when the ensemble members are very different from each other the future state of the atmosphere may be more uncertain (Wilks, 2005).

    If on any given forecast occasion, the observation o is statistically indistinguishable from any of the ensemble members mi, then clearly the bias is zero since E[mi]=E[oi]. Statistical equivalence of any ensemble member and the observation further implies that

    E[(oem)2]=E[(miem)2], (5)

    where em denotes the ensemble mean. By applying the root square to both sides of Eq. (5), it is easy to see that the left-hand side is the RMSE for the ensemble-mean forecasts, whereas the right-hand side expresses dispersion of the ensemble members mi around the ensemble mean (Wilks, 2005). It is important to realize that Eq. (5) holds only for forecasts from a consistent ensemble, and in particular assumes unbiasedness in the forecasts.

    Two consequences of the ensemble consistency condition are that forecasts from the individual ensemble members (and therefore also the ensemble-mean forecasts) are unbiased, and that the average (over multiple forecast occasions) MSE for the ensemble-mean forecasts should be equal to the average ensemble variance (Wilks, 2005).

    Figure 3 displays the dispersion diagram for 10-m wind speed and 2-m temperature. In this diagram, we have plotted the evolution of the average ensemble spread and RMSE of the ensemble mean as function of lead-time, over all the stations for both parameters. This was done for four configurations: (1) basic DM13 version, (2) DM13 with weights issued from machine learning technique and no restriction on analogs dates, (3) DM13 with daily neighborhood restriction, and (4) AnEn with daily neighborhood restriction and weights issued from machine learning technique.

    Fig  3.  (a) Dispersion diagram for 10-m wind speed (random forest) and (b) 2-m temperature (linear regression). The solid line is the average ensemble spread and the dotted/dashed lines are the root-mean-square error (RMSE) of the ensemble mean. Four configurations are shown: DM13 with all predictors weights equal to one (DM13), DM13 with weights issued from machine learning technique (RF/LR + DM13), DM13 with daily neighborhood (DM13 with restriction), and DM13 with weights issued from machine learning technique and daily neighborhood (RF/LR with restriction).

    For 10-m wind speed, DM13 with random forest weights and DM13 basic configurations seem to be the most statistically consistent since the spread is close enough to RMSE with the minimum RMSE linked to DM13 with random forest configuration. For T2m, DM13 with daily neighborhood restriction is the best configuration in terms of statistical consistency (differences magnitude not exceeding 0.25°C) followed by AnEn configuration with daily neighborhood restriction and linear regression weights.

    For the other parameters, the most statistically consistent configurations are DM13 with XGBoost weights for RH2m (difference below 1%), DM13 with random forest weights for MSLP and SURFP (difference below 0.5 hPa), and DM13 with weights from XGBoost for WD10m (not shown).

    For most parameters, it is found that the good statistical consistency of the AnEn comes from a lower RMSE of the ensemble mean, which is achieved from use of higher-resolution NWP AROME model. In addition, a lower ensemble spread also contributes to the consistency of the forecast system.

    Overall, it is found that configurations with weights issued from machine learning techniques without applying the daily neighborhood restriction are almost the most statistically consistent with the lowest RMSE. A more in-depth assessment of a statistical consistency at a particular season or forecast lead time will be discussed later.

    In this section, the AnEn performance with the best configurations, trained during 2016–2018, is assessed and compared for each of the studied surface weather parameters over 2019, firstly to the operational NWP AROME model outputs and secondly to the AnEn configuration of DM13 where equal weights (= 1) are assigned to predictors and no daily neighborhood restriction is applied. To achieve this, Fig. 4 displays the temporal evolution of the Bias (mean error) and RMSE as function of lead time for T2m, RH2m, MSLP, SURFP, WD10, and WS10. The plotted metrics are the average value over all the 15 airports.

    Fig  4.  Bias and RMSE curves for continuous prediction of the following surface parameters (T2m, RH2m, MSLP, SURFP, WS10m, and WD10m) using the best daily neighborhood restriction and machine learning weighting configuration (presented in dotted blue line), NWP AROME prediction of those parameters are plotted in black solid line, and finally the basic DM13 method is plotted in triangular red line.

    From Fig. 4, the performance of the operational model AROME (black line) shows a high variability, depending on the studied surface weather parameter and lead time. For instance, AROME predicts RH2m with moist bias not exceeding 3% associated with an RMSE of 10%. While AROME shows a cold bias of about −0.5°C and an RMSE of 1.5°C for T2m. For WS10m and WD10m, AROME Bias is below 0.25 m s−1 and 3 degrees respectively. For barometric fields, both MSLP and SURFP have a bias below 1 hPa and an RMSE of 2 hPa. These scores are similar to the findings of Seity et al. (2011).

    Comparing the performances of the basic DM13 and those of the modified version configurations, the latter displays significant improvement of RMSE compared to DM13 for MSLP, SURFP, and T2m, while DM13 has the best RMSE for RH2m. Regarding the Bias, it is found that the basic DM13 has the best score for T2m, MSLP, and SURFP. The modified version shows better Bias only for RH2m; however, both methods show similar Bias values for WD10m and WS10m, where they outperform slightly each other for different lead times.

    Comparing AnEn configurations and AROME performances, it is found that AnEn outperforms significantly AROME in terms of Bias during the night and early morning (lead times from 0 to 10 h and from 18 to 23 h) for all parameters. Regarding the RMSE, the best scores of AnEn are found for WS10m and WD10m.

    During the nighttime forecast hours, a reduction varying between 7% and 20% of AROME’s error for WS10m, SURFP, and WD10m is perceived. These improvements in RMSE can be explained by the fact that AnEn ensemble members are constructed mainly from real observations that in 60% cases belong to the daily neighborhood (not shown), which preserves the variation in RMSE magnitude. In addition, machine learning AnEn’s configurations take into account only significant predictors. This avoids using irrelevant predictors that add more error and inaccuracy to analogs selection process. Feature selection techniques often reduce the RMSE of regression coefficients, in particular for weak or noisy predictors (Heinze et al., 2018). However, a slight degradation of the RMSE is found for the thermodynamic surface parameters (RH2m and T2m) in comparison with AROME (Fig. 4). This could be due to the high spatiotemporal variability of these parameters and also to the model forecast error. This might indicate that AnEn needs more predictors representing the lower layer of the atmospheric boundary layer instead of using only the main surface weather parameters issued from the SYNOP messages.

    From a seasonal dependency point of view, it is found that the modified version AnEn DM13 yield improvement of Bias and RMSE in winter and spring during the daytime forecast hours (from 10 to 18 h) while slight worsenings are found for both metrics in summer and autumn (not shown). This is mainly due to the limited ability of AROME in predicting the rapid convective mesoscale situations during these months (Seity et al., 2011). In the next section, the spatial distribution of AnEn performance is investigated to assess any dependency of region and season.

    To assess the global performance of AnEn over the study domain, the Bias and RMSE gain/loss [Eq. (6)], in comparison with AROME, have been calculated for each airport over all the lead times and are plotted spatially (RH2m as an example in Fig. 5) for DM13 and the best configuration of the modified version of DM13.

    Fig  5.  Bias and RMSE gain of RH2m at each airport for DM13 and AnEn’s best configuration. The same metrics (Bias and RMSE) for the operational AROME NWP model at each station are also plotted as a benchmark.
    Gain/Loss={sign(Scorenwp)×sign(Scoreanen)×|Scorenwp||Scoreanen||Scorenwp|,Scorenwp0sign(Scoreanen)×0.01|Scoreanen|0.01,Scorenwp=0, (6)

    where Scoreanen and Scorenwp are relatively the scores for AnEn and NWP.

    Indeed, when the NWP score is perfect (= 0), this score is converted to 0.01 to avoid infinite loss value.

    For 2-m relative humidity (Fig. 5), AROME model draws a positive bias over coastal areas (up to 10%), and negative bias for two airports in the northeastern part, while the RH2m bias is very weak in the other airports. With regard to AROME, AnEn yielded an important improvement by reducing RH2m Bias by up to 50% for almost all the airports while it reduced the RH2m RMSE by about 25%. It should be noted that AnEn degrades the AROME Bias in six airports in the north part of Morocco; most of these stations already have an almost perfect Bias by AROME.

    Similarly, the spatial AnEn performance analysis for the other surface parameters points out that AROME shows a negative Bias for T2m especially over coastal areas (not shown), while AnEn highlights an improvement up to 50% in Bias and 25% in RMSE for most of the airports from different regions (Mediterranean, Atlantic coastal areas, southern region, interior, and eastern region). For the rest of airports, AnEn converts negative Bias to positive Bias with magnitude not exceeding 1.5°C.

    Besides, AROME highlights a negative bias for mean sea level pressure and surface pressure all over Mediterranean and Atlantic coastal areas and a weak bias far inland. AnEn reduced pressure bias and RMSE magnitude concurrently by 50% for most airports, except a few interior airports where AnEn degrades these metrics. AnEn best configuration reduced Bias and RMSE of wind speed by more than 40% of its magnitude, this finding is generalized for all airports.

    For wind direction, AROME underlines very weak Bias in general except for few interior airports where Bias exceeds 9 degrees, while AnEn reduced Bias magnitude in those areas to 2.5 degree or by 50% and preserved slightly bias nullity all over the rest. The RMSE of wind direction was reduced by AnEn up to 60%. For wind components, AROME represents a negative Bias over all airports (range from −20 to −5 m s−1), AnEn lower Bias magnitude by 70%. In addition, RMSE error was reduced by 75% for 10-m wind components.

    In Figs. 6 and 7, we have plotted the seasonal spatial distribution of gain/loss in Bias and RMSE for T2m, using DM13 random forest configuration. It is seen clearly that AnEn performances are seasonally dependent. Indeed, Bias gain exceeds 50% for most of the airports especially in autumn and summer; while a general worsening is found in spring, one can remark that Bias gain reached 70% for some locations on the Atlantic side. Regarding the RMSE, seasonal gain for T2m is around 20% for at least half of the studied airports except for spring.

    Fig  6.  Bias gain/loss for T2m for all seasons using no daily neighborhood restriction and random forest weights configuration.
    Fig  7.  RMSE gain/loss for T2m for all seasons using no daily neighborhood restriction and random forest weights configuration.

    Overall, the gain/loss behavior for Bias and RMSE is different from one parameter to another and is highly dependent on season and location. In this context, Bias gain exceeds 50% for 70% of parameters (not shown). Slight degradations are observed in the airports with good seasonal Bias.

    A limitation of AnEn is a slight degradation of seasonal RMSE for some tested configurations and parameters (not shown). This RMSE loss is mainly observed in the Mediterranean and eastern airports (Tanger, Al Hoceima, and Oujda). It is also found that neighborhood restrictions showed valuable improvements for some parameters (T2m, RH2m, WS10m) and contributed to increase the RMSE and Bias gain for different locations and seasons for those parameters.

    As mentioned before, AnEn leverages a single deterministic NWP model to produce probabilistic forecasts. Thus, it is highly recommended to use the same model. In our case, only the lateral boundary conditions have been changed in 2020.

    First, the impact of this change on the AROME performance itself is assessed. It is found that the new boundary condition coupling used in 2020 softly impacted AROME’s performance for most of the parameters (10-m wind speed is shown as an example in Fig. 8). Indeed, improvements in RMSE (not shown) and Bias are noticeable mainly for synoptic dynamical parameters such as MSLP, SURFP, WS10m, and 10-m wind components. On the other hand, AROME’s performances are of no significant change for RH2m, T2m, and WD10m.

    Fig  8.  Wind speed Bias over 2020 for all lead times and all the possible AnEn configurations.

    However, AnEn performance was not affected by the boundary conditions change. Indeed AnEn still outperforms AROME spatially, temporally, and seasonally for most of the surface parameters and locations. AnEn still shows best Bias and RMSE during nights and mornings over 2020 for most of the surface parameters. A finding to underline is that AnEn Bias gets improved during day lead times where it overtakes AROME for all parameters except MSLP and SURFP. On the other hand, AnEn RMSE remains the best overall. It is to mention that AnEn best configurations of 2019 are not preserved for 2020. For instance for WS10m, the best AnEn configuration during 2019 was DM13 + RF while it is DM13 + XGB for 2020.

    In this study, it is demonstrated that the best configurations with multiple criteria of AnEn yield important improvement in surface weather parameters forecasting but it still has some shortcomings.

    Indeed, the perfect analogy is far from existing, but identifying close enough situations leading to similar effects is still possible. Then, the relevance of analogy and thus analogs forecasting quality is tightly affected by the three major following factors: i) the process of skillful analogs selection in the training data, which depend on the similarity metric, predictors weighting and selection, temporal window around the target lead time, and the number of members; ii) the target surface weather parameter and its predictability by data-driven forecasting techniques; and iii) the used NWP model error in the training data and its ability to forecast rare events and some mesoscale phenomena (Zhao and Giannakis, 2016).

    In this research, we tackled the main surface weather parameters issued from SYNOP messages. While AnEn improves the synoptic scale parameter forecasting (MSLP, SURFP, WS10, and WD10) by reducing both Bias and RMSE, it is found that AnEn improves slightly the Bias but degrades the RMSE for the thermodynamic surface parameters (RH2m and T2m) during the daytime lead times and also spatially in some airports. This could be overcome by adding predictors representing the lower layer of the atmospheric boundary layer in addition to parameters that describe the atmospheric circulation such as geopotential fields in many levels (Duband, 1981; Guilbaud, 1997) or new sets of predictors at different pressure levels (Z500, TPW850, etc.) (Horton et al., 2012). In fact, even for two locations that are close to each other but subject to different critical atmospheric conditions, the selection of the best predictors can vary. Thus, the method needs to be adapted to local conditions, available data, and the size of the region of interest. In this framework, three machine learning techniques (linear regression, random forest, and extreme gradient boosting) have been used in this study to find the optimal weights for the physically meaningful predictors. This step of AnEn forecasting process aims to maximize the useful information and reduce noise. The optimal importance of predictors however varies from region to another, a season to another, and along with the leading atmospheric process. Indeed, Junk et al. (2015) stated that optimized predictor weights are highly affected by terrain complexity and atmospheric stratification. In our case, using machine learning technique to optimize the predictor weights leads to an improvement in Bias and RMSE exceeding 50% while the gain reached 21% in Gensler et al. (2016), 20% in Junk et al. (2015), and 44% in Wang et al. (2019).

    One key aspect in the analogy is the seasonal preselection of the analogs. This preselection is implemented in this study as a moving selection of ±15 days centered around the target date for every year of the archive and time window of ±1 h around the forecast hour since the hourly forecasts are used here and are issued from a high-resolution mesoscale operational model. However, it is found that seasonal preselection yields worsening of AnEn performance in some airports. Indeed, for one target day, the sampling (31 days/yr × 3 yr of training = 93 days) might be inadequate to retrieve the skillful analogs due to the missing observed values or to the occurrence of rare events in this temporal window. Thus, it is concluded that this approach requires a very long archive that is why no restriction on target date neighboring performs better. Indeed, an analysis of the selected analogs position from the target has been performed and it is found that for some cases, many analogs get outside the daily neighborhood window (±15 days). Hence, it would be very beneficial to extend that window. This is in line with the finding of many previous studies for climate purpose that uses daily neighborhood criterion to detect relevant analogs (Bontron, 2004; Horton et al., 2012; Ben Daoud et al., 2016). In ensemble forecasting, the number of members is a parameter of higher relevance. Hence, finding the optimal number of members (analogs) that improves AnEn performances is highly demanded (Horton, 2019; Li et al., 2020). Furthermore, beyond topography, a more precise distinction between spatial areas (urban and non urban for example) can enhance AnEn performance interpretation and point out the ability of using such methods in highly dense urban cities (Li et al., 2020). In addition, some research studies used a large forecast range to investigate the potential of AnEn to capture well the diurnal cycle of the weather surface parameters such as T2m and WS10m (Wang et al., 2019).

    This study presents a new application of the analog ensemble method to improve surface weather parameters prediction over 15 airports of Morocco during 5-yr period (2016–2020), from the non-hydrostatic mesoscale operational model AROME hourly forecasts and observations issued of SYNOP messages. The analog ensemble method application is extended for the first time to include eight main weather surface parameters (T2m, RH2m, MSLP, SURFP, WS10m, WD10m, ZW10m, and MW10m). Seasonal impact was considered in the analogs search process by applying daily neighborhood restriction, in a way that all the analogs have the same season as the current forecast. Best analogs for AnEn are searched by using the studied parameters as predictors, given optimized weights that are calculated as a normalized feature importance coefficients issued from machine learning techniques (linear regression, random forest, and XGBoost) over the training period (2015–2018). Verification of the performance of AnEn was carried out mainly over 2019. An additional verification over 2020 also was held to assess AnEn sensitivity to lateral boundary conditions change.

    The results from the spatial and temporal scores analysis showed that AnEn best configurations produce notably lower Bias and RMSE compared to AROME during night lead times, especially for large scale synoptic parameters (SURFP, MSLP, WD10m, and WS10m). However during day lead times, AnEn shows some limitations, in particular for thermodynamic surface parameters (T2m and RH2m). This is mainly due to the high spatiotemporal variability of these parameters and also to the high RMSE of the ensemble mean and also to higher ensemble spread in some cases during daytime hours. Similar results are reached by AnEn also for most airports, indicated by spatially averaged Bias and RMSE reduction up to 50% and 30% depending on the season. Despite the advantage of lower computational cost, seasonal preselection yields a performance degradation in some airports almost due to the weakness of sampling adequacy.

    According to World Meteorological Organization (WMO), the wind speed value is rounded to the nearest integer in the SYNOP messages. Then, the accuracy of the wind measurement is 1 m s−1. Similarly, the wind direction is coded in a wind rose with 36 directions. This induces an uncertainty of about 10 degrees. The 2-m relative humidity is also rounded to the nearest integer. Consequently, this leads to an accuracy of 1%. All these observation error sources impact the assessment of the AnEn performance while using the continuous verification scores such as Bias and RMSE. One way to overcome these limitations is using the precise observed values from the automatic weather stations that have become an increasingly prominent part of meteorological observation networks over the last 20–30 years and most or all synoptic observations are now automated in some countries.

    The results reported herein can be further improved with a longer training dataset, by extending existing training datasets to consider neighboring locations while searching analogs, exploring further similarity metrics, and by adding more predictors from lower layers of the atmospheric boundary layer or parameters that describe the atmospheric circulation predictors.

    We would like to thank reviewers for taking the time and effort necessary to review the manuscript. We sincerely appreciate all valuable comments and suggestions, which helped us to improve the quality of the manuscript.

  • Fig.  2.   Analog ensemble algorithm steps with day neighborhood and search restrictions.

    Fig.  1.   Map showing orography in meters and the position of the synoptic meteorological stations used in this study, mainly located in airports, and being operational along the whole day (24 h).

    Fig.  3.   (a) Dispersion diagram for 10-m wind speed (random forest) and (b) 2-m temperature (linear regression). The solid line is the average ensemble spread and the dotted/dashed lines are the root-mean-square error (RMSE) of the ensemble mean. Four configurations are shown: DM13 with all predictors weights equal to one (DM13), DM13 with weights issued from machine learning technique (RF/LR + DM13), DM13 with daily neighborhood (DM13 with restriction), and DM13 with weights issued from machine learning technique and daily neighborhood (RF/LR with restriction).

    Fig.  4.   Bias and RMSE curves for continuous prediction of the following surface parameters (T2m, RH2m, MSLP, SURFP, WS10m, and WD10m) using the best daily neighborhood restriction and machine learning weighting configuration (presented in dotted blue line), NWP AROME prediction of those parameters are plotted in black solid line, and finally the basic DM13 method is plotted in triangular red line.

    Fig.  5.   Bias and RMSE gain of RH2m at each airport for DM13 and AnEn’s best configuration. The same metrics (Bias and RMSE) for the operational AROME NWP model at each station are also plotted as a benchmark.

    Fig.  6.   Bias gain/loss for T2m for all seasons using no daily neighborhood restriction and random forest weights configuration.

    Fig.  7.   RMSE gain/loss for T2m for all seasons using no daily neighborhood restriction and random forest weights configuration.

    Fig.  8.   Wind speed Bias over 2020 for all lead times and all the possible AnEn configurations.

    Table  1   Brief description of the machine learning techniques used for computing predictors weights

    MethodDefinition
    Linear regressionIn machine learning literature, regression points out to determine linear relations between the independent and dependent variables recursively via minimization of errors in predictions made by the linear model (Hope, 2020)
    Decision treesDecision trees are a powerful non-parametric supervised machine learning algorithm for classification and regression. The main idea of the algorithm is to best split the dataset into smaller and smaller subsets to predict the target value. The splitting process at each node continues until no further gain can be made or a preset rule is reached. Quality of splits is measured by impurity criteria; for regression one uses variance or mean square error (MSE) reduction to decide to split a node in two or more subnodes. Suppose we are developing a binary tree, the algorithm will first pick a value, and split the data into two subsets. For each subset, it will calculate the MSE separately. The tree chooses the value which results in the smallest MSE value
    Decision trees based techniques (random forest and XGBoost)These techniques merge several decision tree models named estimators to produce an optimum predictive model. This empowers better performances in contrast to a single model. Those methods can be split into two main algorithm families:
    • Boosting and bagging techniques like XGBoost (Chen and Guestrin, 2016) that uses gradient descent minimization on decision trees to create multiple models combined recursively while correcting previous ones to generate a final optimal model. XGBoost is very efficient in terms of usage of computational resources and processing speed (Jhaveri et al., 2019).
    • Random techniques like random forest that stems from decision trees and uses randomness to optimize accuracy by creating a forest of decision trees and merges them to form a more accurate model (Breiman, 2001; Jhaveri et al., 2019)
    Download: Download as CSV

    Table  2   Experimental setup with more details about the used configurations

    ParameterExperiment 1Experiment 2
    Training period2016–20182016–2019
    Verification period20192020
    Boundary conditions changeNoYes
    Location15 synoptic airports of Morocco
    NWP modelAROME 41 cycle, 2.5 km, 90 VLs.
    Ensemble prediction system (EPS) techniqueAnalog ensemble
    Dataset24 hourly observations and forecasts
    FeatureT2m, RH2m, WS10m, WD10m, ZW10m, MW10m, SURFP, and MSLP
    Weighting strategyNo weight/random forest/linear regression/XGBoost
    Time window±1 h
    Day window±15 days
    ConfigurationDM13/DM13 with restriction
    Download: Download as CSV

    Table  3   Percentage of occurrence of possible weights values for each predictor for T2m as predictand

    VariableWeighting value
    0 0.10.20.30.4 0.5 0.6 0.7 0.80.91
    T2m (%)19.4 9.76.5 3.212.922.625.8
    RH2m (%)58.141.9
    WS10m (%)83.916.1
    MSLP (%)51.619.43.29.716.1
    SURFP (%)61.3 6.56.59.716.1
    WD10m (%)83.916.1
    MW10m (%)74.222.63.2
    ZW10m (%)7129
    Download: Download as CSV
  • Alessandrini, S., L. Delle Monache, S. Sperati, et al., 2015: An analog ensemble for short-term probabilistic solar power forecast. Appl. Energy, 157, 95–110. doi: 10.1016/j.apenergy.2015.08.011
    Alessandrini, S., L. Delle Monache, C. M. Rozoff, et al., 2018: Probabilistic prediction of tropical cyclone intensity with an analog ensemble. Mon. Wea. Rev., 146, 1723–1744. doi: 10.1175/MWR-D-17-0314.1
    Ben Daoud, A., E. Sauquet, G. Bontron, et al., 2016: Daily quantitative precipitation forecasts based on the analogue method: Improvements and application to a French large river basin. Atmos. Res., 169, 147–159. doi: 10.1016/j.atmosres.2015.09.015
    Bontron, G., 2004: Prévision quantitative des précipitations: Adaptation probabiliste par recherche d’analogues. Utilisation des Réanalyses NCEP/NCAR et application aux précipitations du Sud-Est de la France. Ph.D. dissertation, Institut National Polytechnique Grenoble (INPG), Grenoble, 289 pp.
    Bousquet, O., D. Barbary, S. Bielli, et al., 2020: An evaluation of tropical cyclone forecast in the Southwest Indian Ocean basin with AROME-Indian Ocean convection-permitting numerical weather predicting system. Atmos. Sci. Lett., 21, e950. doi: 10.1002/asl.950
    Bouttier, F., L. Raynaud, O. Nuissier, et al., 2016: Sensitivity of the AROME ensemble to initial and surface perturbations during HyMeX. Quart. J. Roy. Meteor. Soc., 142, 390–403. doi: 10.1002/qj.2622
    Breiman, L., 2001: Random forests. Machine Learning, 45, 5–32. doi: 10.1023/A:1010933404324
    Brousseau, P., Y. Seity, D. Ricard, et al., 2016: Improvement of the forecast of convective activity from the AROME-France system. Quart. J. Roy. Meteor. Soc., 142, 2231–2243. doi: 10.1002/qj.2822
    Bubnová, R., G. Hello, P. Bénard, et al., 1995: Integration of the fully elastic equations cast in the hydrostatic pressure terrain-following coordinate in the framework of the ARPEGE/Aladin NWP system. Mon. Wea. Rev., 123, 515–535. doi: 10.1175/1520-0493(1995)123<0515:IOTFEE>2.0.CO;2
    Buizza, R., 2008: The value of probabilistic prediction. Atmos. Sci. Lett., 9, 36–42. doi: 10.1002/asl.170
    Buizza, R., and D. Richardson, 2017: 25 years of ensemble forecasting at ECMWF. ECMWF Newsletter No. 153, ECMWF, Reading, 20–31.
    Buizza, R., P. L. Houtekamer, G. Pellerin, et al., 2005: A comparison of the ECMWF, MSC, and NCEP global ensemble prediction systems. Mon. Wea. Rev., 133, 1076–1097. doi: 10.1175/MWR2905.1
    Buizza, R., M. Leutbecher, and L. Isaksen, 2008: Potential use of an ensemble of analyses in the ECMWF ensemble prediction system. Quart. J. Roy. Meteor. Soc., 134, 2051–2066. doi: 10.1002/qj.346
    Chen, T. Q., and C. Guestrin, 2016: XGBoost: A scalable tree boosting system. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, San Francisco, 785–794, doi: 10.1145/2939672.2939785.
    Courtier, P., C. Freydier, J.-F. Geleyn, et al., 1991: The Arpege project at Météo-France. Seminar on Numerical Methods in Atmospheric Models, ECMWF, Shinfield Park, Reading, 193–231.
    Davò, F., S. Alessandrini, S. Sperati, et al., 2016: Post-processing techniques and principal component analysis for regional wind power and solar irradiance forecasting. Solar Energy, 134, 327–338. doi: 10.1016/j.solener.2016.04.049
    Degrauwe, D., Y. Seity, F. Bouyssel, et al., 2016: Generalization and application of the flux-conservative thermodynamic equations in the AROME model of the ALADIN system. Geosci. Model Dev., 9, 2129–2142. doi: 10.5194/gmd-9-2129-2016
    Delle Monache, L., 2015: Probabilistic predictions and downscaling with an analog ensemble for weather, renewable energy, air quality, and hurricane intensity. AGU Fall Meeting Abstracts, American Geophysical Union, A21E-0185.
    Delle Monache, L., F. A. Eckel, D. L. Rife, et al., 2013: Probabilistic weather prediction with an analog ensemble. Mon. Wea. Rev., 141, 3498–3516. doi: 10.1175/MWR-D-12-00281.1
    Delle Monache, L., S. Alessandrini, I. Djalalova, et al., 2018: Air quality predictions with an analog ensemble. Atmos. Chem. Phys. Discuss., 1–36. doi: 10.5194/acp-2017-1214
    Djalalova, I., L. Delle Monache, and J. Wilczak, 2015: PM2.5 analog forecast and Kalman filter post-processing for the Community Multiscale Air Quality (CMAQ) model. Atmos. Environ., 108, 76–87. doi: 10.1016/j.atmosenv.2015.02.021
    Duband, D., 1981: Prévision spatiale des hauteurs de précipitations journalières. La Houille Blanche, 67, 497–512. doi: 10.1051/lhb/1981046
    Eckel, F. A., and C. F. Mass, 2005: Aspects of effective mesoscale, short-range ensemble forecasting. Wea. Forecasting, 20, 328–350. doi: 10.1175/WAF843.1
    Gensler, A., B. Sick, and V. Pankraz, 2016: An analog ensemble-based similarity search technique for solar power forecasting. 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC), IEEE, Budapest, 2850–2857, doi: 10.1109/SMC.2016.7844672.
    Gong, Z. Q., S. F. Li, P. Hu, et al., 2016: Dynamic-analogue correction of the decadal change of East Asian summer precipitation in the late 1990s. J. Meteor. Res., 30, 341–355. doi: 10.1007/s13351-016-5220-1
    Guilbaud, S., 1997: Prévision quantitative des précipitations journalières par une méthode statistico-dynamique de recherche d’analogues - Application à des bassins du pourtour Méditerranéen. Ph.D. dissertation, Institut National Polytechnique de Grenoble, Grenoble, 387 pp.
    Hamill, T. M., and J. S. Whitaker, 2006: Probabilistic quantitative precipitation forecasts based on reforecast analogs: Theory and application. Mon. Wea. Rev., 134, 3209–3229. doi: 10.1175/MWR3237.1
    Hamill, T. M., S. L. Mullen, C. Snyder, et al., 2000: Ensemble forecasting in the short to medium range: Report from a workshop. Bull. Amer. Meteor. Soc., 81, 2653–2680. doi: 10.1175/1520-0477(2000)081<2653:EFITST>2.3.CO;2
    Hamill, T. M., J. S. Whitaker, D. T. Kleist, et al., 2011: Predictions of 2010’s tropical cyclones using the GFS and ensemble-based data assimilation methods. Mon. Wea. Rev., 139, 3243–3247. doi: 10.1175/MWR-D-11-00079.1
    Hansen, B., 2007: A fuzzy logic–based analog forecasting system for ceiling and visibility. Wea. Forecasting, 22, 1319–1330. doi: 10.1175/2007WAF2006017.1
    Hdidou, F. Z., S. Mordane, P. Moll, et al., 2020: Impact of the variational assimilation of ground-based GNSS zenith total delay into AROME-Morocco model. Tellus A: Dynamic Meteorology and Oceanography, 72, 1707854. doi: 10.1080/16000870.2019.1707854
    Heinze, G., C. Wallisch, and D. Dunkler, 2018: Variable selection—A review and recommendations for the practicing statistician. Biometrical Journal, 60, 431–449. doi: 10.1002/bimj.201700067
    Hope, T. M. H., 2020: Linear regression. Machine Learning: Methods and Applications to Brain Disorders. Academic Press, London, 67–81.
    Horton, P., 2019: AtmoSwing: Analog technique model for statistical weather forecastING and downscalING (v2.1.0). Geosci. Model Dev., 12, 2915–2940. doi: 10.5194/gmd-12-2915-2019
    Horton, P., M. Jaboyedoff, R. Metzger, et al., 2012: Spatial relationship between the atmospheric circulation and the precipitation measured in the western Swiss Alps by means of the analogue method. Nat. Hazards Earth Syst. Sci., 12, 777–784. doi: 10.5194/nhess-12-777-2012
    Horton, P., C. Obled, and M. Jaboyedoff, 2017: The analogue method for precipitation prediction: finding better analogue situations at a sub-daily time step. Hydrol. Earth Syst. Sci., 21, 3307–3323. doi: 10.5194/hess-21-3307-2017
    Jhaveri, S., I. Khedkar, Y. Kantharia, et al., 2019: Success prediction using random forest, CatBoost, XGBoost and AdaBoost for Kickstarter campaigns. 2019 3rd International Conference on Computing Methodologies and Communication (ICCMC), IEEE, Erode, 1170–1173, doi: 10.1109/ICCMC.2019.8819828.
    Jolliffe, I. T., and D. B. Stephenson, 2003: Forecast Verification: A Practitioner’s Guide in Atmospheric Science. Wiley, New York, 240 pp.
    Junk, C., L. Delle Monache, S. Alessandrini, et al., 2015: Predictor-weighting strategies for probabilistic wind power forecasting with an analog ensemble. Meteor. Z., 24, 361–379. doi: 10.1127/metz/2015/0659
    Knippertz, P., M. Christoph, and P. Speth, 2003: Long-term precipitation variability in Morocco and the link to the large-scale circulation in recent and future climates. Meteor. Atmos. Phys., 83, 67–88. doi: 10.1007/s00703-002-0561-y
    Lafore, J. P., J. Stein, N. Asencio, et al., 1998: The Meso-NH Atmospheric Simulation System. Part I: adiabatic formulation and control simulations. Ann. Geophys., 16, 90–109. doi: 10.1007/s00585-997-0090-6
    Lagarias, J. C., J. A. Reeds, M. H. Wright, et al., 1998: Convergence properties of the Nelder–Mead simplex method in low dimensions. SIAM J. Optim., 9, 112–147. doi: 10.1137/S1052623496303470
    Leith, C. E., 1974: Theoretical skill of Monte Carlo forecasts. Mon. Wea. Rev., 102, 409–418. doi: 10.1175/1520-0493(1974)102<0409:TSOMCF>2.0.CO;2
    Lguensat, R., P. Tandeo, P. Ailliot, et al., 2017: The analog data assimilation. Mon. Wea. Rev., 145, 4093–4107. doi: 10.1175/MWR-D-16-0441.1
    Li, Z.-M., X.-J. Zhao, Z.-B. Sun, et al., 2020: Research on the interpretation and correction of numerical ozone forecast based on Analog Ensemble. China Environ. Sci., 40, 475–484. (in Chinese) doi: 10.19674/j.cnki.issn1000-6923.2020.0102
    Molteni, F., R. Buizza, T. N. Palmer, et al., 1996: The ECMWF Ensemble Prediction System: Methodology and validation. Quart. J. Roy. Meteor. Soc., 122, 73–119. doi: 10.1002/qj.49712252905
    Orrell, D., L. Smith, J. Barkmeijer, et al., 2001: Model error in weather forecasting. Nonlin. Processes Geophys., 8, 357–371. doi: 10.5194/npg-8-357-2001
    PaiMazumder, D., and N. Mölders, 2009: Theoretical assessment of uncertainty in regional averages due to network density and design. J. Appl. Meteor. Climatol., 48, 1643–1666. doi: 10.1175/2009JAMC2022.1
    Palmer, T. N., 2001: A nonlinear dynamical perspective on model error: A proposal for non-local stochastic-dynamic parametrization in weather and climate prediction models. Quart. J. Roy. Meteor. Soc., 127, 279–304. doi: 10.1002/qj.49712757202
    Riordan, D., and B. K. Hansen, 2002: A fuzzy case-based system for weather prediction. Eng. Intell. Syst., 10, 139–146.
    Saaty, R. W., 1987: The analytic hierarchy process—what it is and how it is used. Mathematical Modelling, 9, 161–176. doi: 10.1016/0270-0255(87)90473-8
    Sahlaoui, Z., S. Mordane, E. Wattrelot, et al., 2020: Improving heavy rainfall forecasts by assimilating surface precipitation in the convective scale model AROME: A case study of the Mediterranean event of November 4, 2017. Meteor. Appl., 27, e1860. doi: 10.1002/met.1860
    Seity, Y., P. Brousseau, S. Malardel, et al., 2011: The AROME-France convective-scale operational model. Mon. Wea. Rev., 139, 976–991. doi: 10.1175/2010MWR3425.1
    Toth, Z., 2001: Meeting summary: Ensemble forecasting in WRF. Bull. Amer. Meteor. Soc., 82, 695–697. doi: 10.1175/1520-0477(2001)082<0681:MSGMIW>2.3.CO;2
    Toth, Z., and E. Kalnay, 1993: Ensemble forecasting at NMC: The generation of perturbations. Bull. Amer. Meteor. Soc., 74, 2317–2330. doi: 10.1175/1520-0477(1993)074<2317:EFANTG>2.0.CO;2
    Toth, Z., and E. Kalnay, 1997: Ensemble forecasting at NCEP and the breeding method. Mon. Wea. Rev., 125, 3297–3319. doi: 10.1175/1520-0493(1997)125<3297:EFANAT>2.0.CO;2
    Tuba, Z., and Z. Bottyán, 2018: Fuzzy logic-based analogue forecasting and hybrid modelling of horizontal visibility. Meteor. Atmos. Phys., 130, 265–277. doi: 10.1007/s00703-017-0513-1
    Vié, B., O. Nuissier, and V. Ducrocq, 2011: Cloud-resolving ensemble simulations of Mediterranean heavy precipitating events: Uncertainty on initial conditions and lateral boundary conditions. Mon. Wea. Rev., 139, 403–423. doi: 10.1175/2010MWR3487.1
    Wang, Z. W., M. Chen, L. Delle Monache, et al., 2019: Application of analog ensemble method to surface temperature and wind speed prediction in Beijing area. Acta Meteor. Sinica, 77, 869–884. (in Chinese) doi: 10.11676/qxxb2019.044
    Whitaker, J. S., T. M. Hamill, X. Wei, et al., 2008: Ensemble data assimilation with the NCEP Global Forecast System. Mon. Wea. Rev., 136, 463–482. doi: 10.1175/2007MWR2018.1
    Wilks, D. S., 2005: Statistical Methods in the Atmospheric Sciences. 2nd ed. Academic Press, London, 648 pp.
    Yu, H. P., J. P. Huang, W. J. Li, et al., 2014: Development of the analogue-dynamical method for error correction of numerical forecasts. J. Meteor. Res., 28, 934–947. doi: 10.1007/s13351-014-4077-4
    Zhao, Z. Z., and D. Giannakis, 2016: Analog forecasting with dynamics-adapted kernels. Nonlinearity, 29, 2888–2939. doi: 10.1088/0951-7715/29/9/2888
    Zhou, X. Q., Y. J. Zhu, D. C. Hou, et al., 2017: Performance of the new NCEP global ensemble forecast system in a parallel experiment. Wea. Forecasting, 32, 1989–2004. doi: 10.1175/WAF-D-17-0023.1
    Zhu, Y. J., 2005: Ensemble forecast: A new approach to uncertainty and predictability. Adv. Atmos. Sci., 22, 781–788. doi: 10.1007/BF02918678
  • Related Articles

  • Other Related Supplements

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return