diff --git a/CHANGELOG.md b/CHANGELOG.md index 286fd915..19160c59 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -8,8 +8,12 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 Please add your functional changes to the appropriate section in the PR. Keep it human-readable, your future self will thank you! -## [Unreleased](https://github.com/ecmwf/anemoi-training/compare/0.3.1...HEAD) +## [Unreleased](https://github.com/ecmwf/anemoi-training/compare/0.3.2...HEAD) + +## [0.3.2 - Multiple Fixes, Checkpoint updates, Stretched-grid/LAM updates](https://github.com/ecmwf/anemoi-training/compare/0.3.1...0.3.2) - 2024-12-19 + ### Fixed + - Not update NaN-weight-mask for loss function when using remapper and no imputer [#178](https://github.com/ecmwf/anemoi-training/pull/178) - Dont crash when using the profiler if certain env vars arent set [#180](https://github.com/ecmwf/anemoi-training/pull/180) - Remove saving of metadata to training checkpoint [#57](https://github.com/ecmwf/anemoi-training/pull/190) @@ -19,8 +23,10 @@ Keep it human-readable, your future self will thank you! - Identify stretched grid models based on graph rather than configuration file [#204](https://github.com/ecmwf/anemoi-training/pull/204) ### Added + - Introduce variable to configure: transfer_learning -> bool, True if loading checkpoint in a transfer learning setting. -- TRANSFER LEARNING: enabled new functionality. You can now load checkpoints from different models and different training runs. +- + TRANSFER LEARNING: enabled new functionality. You can now load checkpoints from different models and different training runs. - Effective batch size: `(config.dataloader.batch_size["training"] * config.hardware.num_gpus_per_node * config.hardware.num_nodes) // config.hardware.num_gpus_per_model`. Used for experiment reproducibility across different computing configurations. - Added a check for the variable sorting on pre-trained/finetuned models [#120](https://github.com/ecmwf/anemoi-training/pull/120) @@ -33,11 +39,13 @@ Keep it human-readable, your future self will thank you! ### Changed ### Removed + - Removed the resolution config entry [#120](https://github.com/ecmwf/anemoi-training/pull/120) ## [0.3.1 - AIFS v0.3 Compatibility](https://github.com/ecmwf/anemoi-training/compare/0.3.0...0.3.1) - 2024-11-28 ### Changed + - Perform full shuffle of training dataset [#153](https://github.com/ecmwf/anemoi-training/pull/153) ### Fixed @@ -47,11 +55,11 @@ Keep it human-readable, your future self will thank you! - Allow histogram and spectrum plot for one variable [#165](https://github.com/ecmwf/anemoi-training/pull/165) ### Added + - Introduce variable to configure (Cosine Annealing) optimizer warm up [#155](https://github.com/ecmwf/anemoi-training/pull/155) - Add reader groups to reduce CPU memory usage and increase dataloader throughput [#76](https://github.com/ecmwf/anemoi-training/pull/76) - Bump `anemoi-graphs` version to 0.4.1 [#159](https://github.com/ecmwf/anemoi-training/pull/159) - ## [0.3.0 - Loss & Callback Refactors](https://github.com/ecmwf/anemoi-training/compare/0.2.2...0.3.0) - 2024-11-14 ### Fixed @@ -60,8 +68,10 @@ Keep it human-readable, your future self will thank you! - Refactored callbacks. [#60](https://github.com/ecmwf/anemoi-training/pulls/60) - Updated docs [#115](https://github.com/ecmwf/anemoi-training/pull/115) - Fix enabling LearningRateMonitor [#119](https://github.com/ecmwf/anemoi-training/pull/119) + - Refactored rollout [#87](https://github.com/ecmwf/anemoi-training/pulls/87) - Enable longer validation rollout than training + - Expand iterables in logging [#91](https://github.com/ecmwf/anemoi-training/pull/91) - Save entire config in mlflow @@ -71,8 +81,10 @@ Keep it human-readable, your future self will thank you! - Included more loss functions and allowed configuration [#70](https://github.com/ecmwf/anemoi-training/pull/70) - Include option to use datashader and optimised asyncronohous callbacks [#102](https://github.com/ecmwf/anemoi-training/pull/102) - Fix that applies the metric_ranges in the post-processed variable space [#116](https://github.com/ecmwf/anemoi-training/pull/116) + - Allow updates to scalars [#137](https://github.com/ecmwf/anemoi-training/pulls/137) - Add without subsetting in ScaleTensor + - Sub-hour datasets [#63](https://github.com/ecmwf/anemoi-training/pull/63) - Add synchronisation workflow [#92](https://github.com/ecmwf/anemoi-training/pull/92) - Feat: Anemoi Profiler compatible with mlflow and using Pytorch (Kineto) Profiler for memory report [38](https://github.com/ecmwf/anemoi-training/pull/38/) @@ -82,7 +94,6 @@ Keep it human-readable, your future self will thank you! - Functionality to change the weight attribute of nodes in the graph at the start of training without re-generating the graph. [#136] (https://github.com/ecmwf/anemoi-training/pull/136) - Custom System monitor for Nvidia and AMD GPUs [#147](https://github.com/ecmwf/anemoi-training/pull/147) - ### Changed - Renamed frequency keys in callbacks configuration. [#118](https://github.com/ecmwf/anemoi-training/pull/118)