HAPPY

Members

This team has chosen to keep its participants anonymous.


Models

Model name

AZN
Number of individuals supporting model development:
1-5
Maximum number of Central Processing Units (CPUs) supporting model development or forecast production:
< 8
Maximum number of Graphics Processing Units (GPUs) supporting model development or forecast production:
< 4
How would you best classify the IT system used for model development or forecast production:
Single node system

Model summary questionnaire for model AZN

Please note that the list below shows all questionnaires submitted for this model.
They are displayed from the most recent to the earliest, covering each 13-week competition period in which the team competed with this model.

Which of the following descriptions best represent the overarching design of your forecasting model?
  • Machine learning-based weather prediction.
What techniques did you use to initialise your model? (For example: data sources and processing of initial conditions)
The model was trained using data from 2020 to 2024, while inference and prediction were performed using data from 2000 to 2024. First, the original ERA5 daily and monthly datasets were interpolated to a 1.5° grid. Weekly data were then generated by averaging the daily data. Based on the prediction date (the Monday three weeks ahead), the input and output datasets were organized into pairs. The input data include: • Monthly forecast data, • Weekly data of predictive variables for the previous 20 weeks, • Weekly upper-level data for the previous 10 weeks, and • Elevation data. The output data consist of daily predictions for seven specific days: Monday, Wednesday, Friday, and Sunday of the third week ahead, and Tuesday, Thursday, and Saturday of the fourth week ahead. The monthly forecast model was trained using monthly data from 1940 to 2024. It takes the previous 15 months of historical data and other surface variables as input and predicts the monthly mean two months ahead. The monthly forecast data used as input correspond to the predicted month of the target date.
If any, what data does your model rely on for real-time forecasting purposes?
TAS: The input includes 20 consecutive weeks of historical daily 2-meter temperature (tas) and 10 consecutive weeks of historical daily geopotential height at 200 hPa, 300 hPa, and 500 hPa. The monthly forecast model (same as for MSLP) requires 15 weeks of historical data for U10, V10, SST, TAS, and MSLP. MSLP: The input includes 20 consecutive weeks of historical daily mean sea level pressure (mslp) and 10 consecutive weeks of historical daily geopotential height at 500 hPa and 850 hPa, specific humidity at 700 hPa, divergence and potential vorticity at 900 hPa. The monthly forecast model (same as for TAS) requires 15 weeks of historical data for U10, V10, SST, TAS, and MSLP. TP: The input includes 20 consecutive weeks of historical daily total precipitation (tp) and 10 consecutive weeks of historical daily geopotential height at 200 hPa, 300 hPa, and 500 hPa, specific humidity at 700 hPa, and cloud cover at 800 hPa. The monthly forecast model requires 15 weeks of historical data for U10, V10, SST, T2M, MSLP, and TP.
What types of datasets were used for model training? (For example: observational datasets, reanalysis data, NWP outputs or satellite data)
All surface and pressure-level meteorological variables were obtained from ERA5, while the elevation data were obtained from ETOPO Global Relief Model.
Please provide an overview of your final ML/AI model architecture (For example: key design features, specific algorithms or frameworks used, and any pre- or post-processing steps)
The overall prediction framework is divided into two sequential stages. First, more than one year of surface variables is used to predict the corresponding monthly values. Next, using the upper-level variables from the previous three-plus months, the monthly predictions, and elevation data, the model forecasts daily values for seven specific days: Monday, Wednesday, Friday, and Sunday of the third week ahead, and Tuesday, Thursday, and Saturday of the fourth week ahead. The daily predictions for the third and fourth weeks are then converted into quintiles based on historical climate quartiles, and the mean of these quintiles is computed for submission. Separate models are trained for each variable: near-surface air temperature (tas), sea-level pressure (mslp), and total precipitation (tp). Monthly Prediction Model: • Input Data: Surface variables from the previous 15 months. • Architecture: A hybrid of Convolutional Neural Network (CNN) and Temporal Convolutional Network (TCN). • Output: Two-month-ahead single-variable predictions for tas, mslp, or tp. Subseasonal Daily Prediction Model: • Input Data: Weekly data of predictive variables for the previous 20 weeks, weekly upper-level variables for the previous 10 weeks, monthly predictions, and elevation data. • Architectures for Near-Surface Air Temperature (tas): Six deep learning architectures were explored. All showed good convergence during training, with the Normalized Anomaly Correlation (NAC) steadily increasing and ultimately exceeding 0.75: 1. Convolutional Neural Network (CNN) 2. CNN combined with Gated Recurrent Unit (CNN + GRU Hybrid Model) 3. Transformer-CNN Hybrid Model 4. Residual blocks combined with the spatial attention mechanism from the Convolutional Block Attention Module (CBAM) 5. Residual Squeeze-and-Excitation network combined with CNN (Residual SE + CNN) 6. Residual U-Net structure • Architectures for Sea-Level Pressure (mslp): Three architectures demonstrated strong convergence during training, with overall ACC steadily increasing and ultimately exceeding 0.7: 1. Residual U-Net 2. Shallow encoder–decoder Residual Network (ResNet) with CBAM-style spatial attention integrated in each residual block to enhance spatial feature extraction, and a 1×1 convolution decoder to output multivariable predictions 3. Transformer-CNN Hybrid Model, which first maps the multi-channel input to feature embeddings via convolution, then models global spatial dependencies using a Transformer encoder, and finally generates spatial predictions for the target variable via convolutional decoding • Architectures for Total Precipitation (tp): Four architectures demonstrated strong convergence during training, with overall ACC steadily increasing and ultimately exceeding 0.7: 1. Residual U-Net 2. SE Multi-Head ResU-Net, which integrates residual blocks with Squeeze-and-Excitation (SE) modules to enhance feature representation, employs a multi-layer encoder–decoder to extract multi-scale spatial features, and uses multi-head convolutional outputs to generate high-resolution spatial predictions for each target variable 3. Spatial Attention ResNet, which separately encodes different groups of meteorological variables via residual blocks with spatial attention, then fuses these features, and uses independent decoders to produce high-resolution spatial predictions for each target variable • Model Output: Daily predictions for seven specific days: Monday, Wednesday, Friday, and Sunday of the third week ahead, and Tuesday, Thursday, and Saturday of the fourth week ahead. Post-Processing: For each model, the daily predictions for the third and fourth weeks are converted into quintiles using historical climate quartiles. The mean of these quintiles is calculated as the model’s quintile prediction. Finally, the quintiles from all models are averaged to generate the final submission. Historical climate quartiles are computed separately for each model.
Have you published or presented any work related to this forecasting model? If yes, could you share references or links?
I presented an oral talk titled “A Machine Learning Model for Subseasonal Prediction: Forecasting Global Surface Temperature, Pressure, and Precipitation” at the Symposium on Applied Mathematics, Artificial Intelligence Methods, and the Complexity of the Earth System (December 26–28, 2025, Beijing) https://csiam.org.cn/1005/202511/2592.html . The symposium was conducted in Chinese, and the associated conference materials are not publicly documented or available online. Looking forward to the day this work can be published.
Before submitting your forecasts to the AI Weather Quest, did you validate your model against observational or independent datasets? If so, how?
No.
Did you face any challenges during model development, and how did you address them?
(1)I encountered was the apparent randomness in model performance. For instance, when I retrained a previously high-performing temperature prediction architecture with an additional year of data, slightly adjusted evaluation metrics, and some minor optimizations, the resulting model unexpectedly performed worse or became unusable. The underlying cause remains unclear. While this did not affect submission, the instability temporarily reduced my confidence in further model optimization. (2)Another challenge was the limited computational capacity of my laptop. Currently, this does not constrain competition submissions or forecast quality. If I wanted to rigorously assess the impact of computational resources on performance, I would consider purchasing a dedicated GPU rather than relying on cloud computing, budget permitting, as local training allows for more efficient experimentation. Nonetheless, I believe that even with my current setup, there remains substantial potential to further improve forecast accuracy.
Are there any limitations to your current model that you aim to address in future iterations?
On the one hand, the selection of input variables was mainly based on my understanding of meteorological processes and was made rather arbitrarily, without conducting any practical analysis or validation. On the other hand, the model architecture was chosen by testing only a few structures: I kept those that could produce forecasts and discarded those that could not, without exploring a sufficiently wide range of architectures. As a result, I still do not clearly understand what role a specific architecture plays in forecasting. Although both issues are relatively easy to start addressing, they require extensive experiments and cannot be resolved in the short term.
Are there any other AI/ML model components or innovations that you wish to highlight?
(1) The success of the monthly forecasting model marked the beginning of all subsequent possibilities. (2) During the DJF submission period, the improvements I made step by step included adding new sub-models for mslp and tp, and optimizing the method for calculating quintile probabilities. (3) Some mistakes in the SON submission questionnaire included: I trained the model using data from 2020–2024 instead of 2000–2024 (of course, I would prefer to use more training data if computational resources allowed, but I forgot that feeding in 25 years of data would cause the system to crash; later I tested with only five years of data, and after finishing the experiment, I completely overlooked the fact that I had used only five years). In addition, the metric in which tas performed well was NAC, not ACC. Since RMSE contains many digits and looks inconvenient, I wanted to use a metric bounded within 1 for easier evaluation. I thought I had been calculating ACC, but in fact it was NAC. (4) After submitting forecasts for a long time, I realized that some of my approaches to data downloading and folder-structure organization were unnecessarily time-consuming and inefficient—I honestly felt like crying at my own stupidity. /(ㄒoㄒ)/~~
Who contributed to the development of this model? Please list all individuals who contributed to this model, along with their specific roles (e.g., data preparation, model architecture, model validation, etc) to acknowledge individual contributions.
This team has chosen to keep its participants anonymous.

Which of the following descriptions best represent the overarching design of your forecasting model?
  • Machine learning-based weather prediction.
What techniques did you use to initialise your model? (For example: data sources and processing of initial conditions)
The model was trained using data from 2000 to 2024. First, the original ERA5 daily and monthly datasets were interpolated to a 1.5° grid. Weekly data were then generated by averaging the daily data. Based on the prediction date (the Monday three weeks ahead), the input and output datasets were organized into pairs. The input data include: Monthly forecast data, Weekly data of predictive variables for the previous 20 weeks, Weekly upper-level data for the previous 10 weeks, and Elevation data. The output data consist of daily predictions for seven specific days: Monday, Wednesday, Friday, and Sunday of the third week ahead, and Tuesday, Thursday, and Saturday of the fourth week ahead. The monthly forecast model was trained using monthly data from 1940 to 2024. It takes the previous 15 months of historical data and other surface variables as input and predicts the monthly mean two months ahead. The monthly forecast data used as input correspond to the predicted month of the target date.
If any, what data does your model rely on for real-time forecasting purposes?
TAS: The input includes 20 consecutive weeks of historical daily 2-meter temperature (tas) and 10 consecutive weeks of historical daily geopotential height at 200 hPa, 300 hPa, and 500 hPa. The monthly forecast model (same as for MSLP) requires 15 weeks of historical data for U10, V10, SST, TAS, and MSLP. MSLP: The input includes 20 consecutive weeks of historical daily mean sea level pressure (mslp) and 10 consecutive weeks of historical daily geopotential height at 500 hPa and 850 hPa, specific humidity at 700 hPa, divergence and potential vorticity at 900 hPa. The monthly forecast model (same as for TAS) requires 15 weeks of historical data for U10, V10, SST, TAS, and MSLP. TP: The input includes 20 consecutive weeks of historical daily total precipitation (tp) and 10 consecutive weeks of historical daily geopotential height at 200 hPa, 300 hPa, and 500 hPa, specific humidity at 700 hPa, and cloud cover at 800 hPa. The monthly forecast model requires 15 weeks of historical data for U10, V10, SST, T2M, MSLP, and TP.
What types of datasets were used for model training? (For example: observational datasets, reanalysis data, NWP outputs or satellite data)
Only ERA5 reanalysis data.
Please provide an overview of your final ML/AI model architecture (For example: key design features, specific algorithms or frameworks used, and any pre- or post-processing steps)
Model Framework The overall prediction framework consists of two stages. First, the monthly values of each variable are predicted using more than one year of surface variable data. Then, using over three months of upper-level variable data, the monthly forecast results, and elevation data, the model predicts daily values for seven specific days: Monday, Wednesday, Friday, and Sunday of the third week ahead, and Tuesday, Thursday, and Saturday of the fourth week ahead. The weekly mean is then calculated from the predicted data for the third and fourth weeks, and the quintile category is determined for submission. Each target variable (TAS, MSLP, and TP) is trained with its own independent model. Monthly Forecast Model Input data: Surface variables from the previous 15 months. Model architecture: A hybrid structure combining a Convolutional Neural Network (CNN) and a Temporal Convolutional Network (TCN). Output data: Single-variable (TAS, MSLP, or TP) forecasts two months ahead. Main Prediction Model Input data: Weekly data of the target variable from the previous 20 weeks, Weekly upper-level data from the previous 10 weeks, Monthly forecast data, and Elevation data. Model architecture: For near-surface air temperature (TAS), six deep learning architectures were tested. All demonstrated stable convergence during training (with overall ACC gradually increasing and reaching above 0.75). The architectures include: Convolutional Neural Network (CNN) CNN + GRU Hybrid Model Transformer–CNN Hybrid Model CBAM-based model combining spatial attention mechanisms and residual blocks Residual Squeeze-and-Excitation + CNN Model Residual U-Net Model In comparison, for mean sea level pressure (MSLP) and total precipitation (TP), only the Residual U-Net architecture achieved stable and effective performance during training. Output prediction: Daily forecasts for seven specific days—Monday, Wednesday, Friday, and Sunday of the third week ahead, and Tuesday, Thursday, and Saturday of the fourth week ahead. Post-processing For each model, the predicted data for the third and fourth weeks are averaged, and the resulting mean is compared with the historical climatological quartiles to determine the quintile category. For TAS, six models were used to generate historical climatological quartiles and corresponding quintile predictions. The final TAS result is obtained by averaging the quintile outputs from all six models. For MSLP and TP, the quintile prediction is derived directly from the single Residual U-Net model.
Have you published or presented any work related to this forecasting model? If yes, could you share references or links?
No, we plan to start writing the related content after gaining better control over the model’s performance.
Before submitting your forecasts to the AI Weather Quest, did you validate your model against observational or independent datasets? If so, how?
No.
Did you face any challenges during model development, and how did you address them?
The model was trained entirely on a gaming laptop. Although the computational power was limited, which constrained both data input volume and model complexity, the model was successfully completed and submitted. The reason for firmly choosing to train on a gaming laptop was twofold: first, it ensured stable computational access throughout the entire competition period; second, I did not have access to large or fast computing resources, so I aimed to develop models optimized for low computational capacity. When training became slow or unresponsive, I simplified the model architecture, reduced the number of input variables, and repeated the process several times until it could run smoothly.
Are there any limitations to your current model that you aim to address in future iterations?
The model is still in its early development stage, and many aspects are being adjusted and optimized. For this submission cycle, it became clear that the conversion from the predicted true values of sea level pressure and precipitation to their quintile categories could be improved by using predicted historical values to calculate climatological means—similar to how it was done for near-surface air temperature. In future work, I plan to refine this conversion process to achieve better evaluation scores, and to carefully analyze the prediction performance to identify issues and further optimize the model.
Are there any other AI/ML model components or innovations that you wish to highlight?
The main motivation and foundation for developing this model came from testing monthly-scale predictions using ERA5 single-level data. By predicting tas one month ahead based on the previous 15 months of historical data, the model achieved excellent RMSE and ACC performance, which gave me strong confidence to proceed with subsequent developments.
Who contributed to the development of this model? Please list all individuals who contributed to this model, along with their specific roles (e.g., data preparation, model architecture, model validation, etc) to acknowledge individual contributions.
This team has chosen to keep its participants anonymous.

Model name

BZN
Number of individuals supporting model development:
1-5
Maximum number of Central Processing Units (CPUs) supporting model development or forecast production:
< 8
Maximum number of Graphics Processing Units (GPUs) supporting model development or forecast production:
< 4
How would you best classify the IT system used for model development or forecast production:
Single node system

Model summary questionnaire for model BZN

Please note that the list below shows all questionnaires submitted for this model.
They are displayed from the most recent to the earliest, covering each 13-week competition period in which the team competed with this model.

Which of the following descriptions best represent the overarching design of your forecasting model?
  • Other: During the DJF submission period, the main update was switching to a different method for computing the AZN quintile probability results. In the subsequent phase, I plan to submit results based on a new forecasting approach.
What techniques did you use to initialise your model? (For example: data sources and processing of initial conditions)
Same as AZN during the DJF period.
If any, what data does your model rely on for real-time forecasting purposes?
Same as AZN during the DJF period.
What types of datasets were used for model training? (For example: observational datasets, reanalysis data, NWP outputs or satellite data)
Same as AZN during the DJF period.
Please provide an overview of your final ML/AI model architecture (For example: key design features, specific algorithms or frameworks used, and any pre- or post-processing steps)
Compared with AZN during the DJF period, I computed the climatological quartiles for each forecast lead time, then derived the quintile probabilities for each lead time. Next, I calculated the weekly mean of the quintile probabilities based on which week each lead time belongs to. The submitted quintile probabilities were obtained by averaging the quintile probabilities across all models. I may conduct some different tests during the submission period.
Have you published or presented any work related to this forecasting model? If yes, could you share references or links?
No
Before submitting your forecasts to the AI Weather Quest, did you validate your model against observational or independent datasets? If so, how?
No
Did you face any challenges during model development, and how did you address them?
No
Are there any limitations to your current model that you aim to address in future iterations?
I wanted to submit results using a new forecasting approach, and I also wanted to obtain a full ranking for the entire MAJ season, so I registered a new model in time (≧▽≦)ゞ!
Are there any other AI/ML model components or innovations that you wish to highlight?
No
Who contributed to the development of this model? Please list all individuals who contributed to this model, along with their specific roles (e.g., data preparation, model architecture, model validation, etc) to acknowledge individual contributions.
This team has chosen to keep its participants anonymous.

Submitted forecast data in previous period(s)

Please note: Submitted forecast data is only publicly available once the evaluation of a full competitive period has been completed. See the competition's full detailed schedule with submitted data publication dates for each period here.

Access forecasts data

Participation

Competition Period

For the selected competition period, the table below shows the variables submitted each week by the respective team.

Week First forecast window: Days 19 to 25 Second forecast window: Days 26 to 32
Near-surface (2m) temperature (tas) Mean sea level pressure (mslp) Precipitation (pr) Near-surface (2m) temperature (tas) Mean sea level pressure (mslp) Precipitation (pr)

This team did not submit any entries to the competition