Quickstart¶
This page provides the fastest way to verify the released benchmark resources. It focuses on evaluation-first usage: preparing the dataset and pretrained checkpoints, and then running the provided evaluation scripts for the released UNet and GAN baselines.
If your goal is to: - inspect the released data, start from the Dataset page - evaluate released checkpoints, follow this page - retrain models or reproduce the generation pipeline, refer to the code repository and detailed documentation pages
1. Prepare dataset and pretrained checkpoints¶
Place the dataset and pretrained models under the project root as:
<project_root>/
Dataset/
radiomaps/
height_maps/
beam_maps/
configs/
sionna_maps/ # optional
Pretrained_Model/
GAN/
...
Unet/
...
The released evaluation scripts assume these default paths.
Required folders for evaluation¶
For evaluating pretrained checkpoints, the required folders are:
Dataset/radiomapsDataset/height_mapsDataset/beam_mapsPretrained_Model/UnetPretrained_Model/GAN
The folder Dataset/sionna_maps is optional and is only needed if you want to reproduce the ray-tracing generation pipeline.
2. Recommended first run¶
For a first verification, the recommended order is:
- run the released UNet evaluation script
- run the released GAN evaluation script
This provides a quick sanity check that:
- dataset paths are correct
- pretrained checkpoints are correctly placed
- the evaluation pipeline runs end-to-end
3. Evaluate pretrained UNet checkpoints¶
Script:
ModelEvaluation_Unet.py
Run:
python ModelEvaluation_Unet.py
The UNet evaluation script uses a built-in configuration class (EvalConfig) and does not require command-line arguments.
Default paths¶
By default, the script uses:
Dataset/radiomapsDataset/height_mapsDataset/beam_mapsPretrained_Model/Unet
Default evaluation settings¶
RANDOM_SEED = 42TRAIN_RATIO = 0.7VAL_RATIO = 0.1TEST_RATIO = 0.2BATCH_SIZE = 64
Outputs¶
Results are saved to:
evaluation_results/
Generated outputs include:
evaluation_summary_dB.csvmetrics_comparison_dB.png{model_name}_visualization.png
Notes¶
- metrics are computed in the dB domain
- building regions and invalid regions are excluded
- SSIM is computed only on valid regions
If you want to change settings¶
ModelEvaluation_Unet.py evaluates a predefined model list from EvalConfig.MODELS.
To change:
- evaluated checkpoints
- dataset paths
- split strategy
- batch size
- model list
please edit the corresponding fields in EvalConfig.
4. Evaluate pretrained GAN checkpoints¶
Script:
ModelEvaluation_GAN.py
Run:
python ModelEvaluation_GAN.py
The GAN evaluation script automatically scans the experiment folders under:
Pretrained_Model/GAN/
and evaluates discovered experiments containing:
best_G.pth
Default paths¶
By default, the script uses:
Dataset/radiomapsDataset/height_mapsDataset/beam_mapsPretrained_Model/GAN
Outputs¶
Typical outputs include:
Pretrained_Model/GAN/evaluation_summary.jsonPretrained_Model/GAN/evaluation_summary.csv
Per-experiment files may also include:
eval_results.jsoneval_visualization.png
Supported released task folders¶
The released GAN evaluator is aligned with task folders such as:
beam_dense_encodingbeam_dense_featurerandom_dense_encodingrandom_dense_featurerandom_sparse_encoding_samples819random_sparse_feature_samples819scene_dense_encodingscene_dense_feature
These are automatically detected from experiment folders containing best_G.pth.
Configuration inference¶
The script infers task settings from one of the following sources:
- predefined
EXPERIMENT_CONFIGS config.json- experiment folder naming rules
Optional notebook-style usage¶
The script can also be used interactively. For example:
results = run_evaluation()
results = run_evaluation(single='random_dense_feature')
results = run_evaluation(visualize=True)
5. Common issues¶
File not found¶
Check that the following folders exist exactly as expected:
Dataset/radiomapsDataset/height_mapsDataset/beam_mapsPretrained_Model/GANPretrained_Model/Unet
GPU selection¶
The UNet evaluation script sets:
os.environ["CUDA_VISIBLE_DEVICES"] = "0"
If you need another GPU, modify this value in the script.
Missing Python packages¶
The evaluation scripts require packages such as:
torchnumpymatplotlibpandasscikit-image
The UNet evaluation script also imports:
seaborn
6. What this quickstart covers¶
This page focuses on evaluation of released checkpoints.
It does not cover in detail:
- retraining the GAN or UNet baselines
- regenerating the dataset assets
- reproducing the complete ray-tracing pipeline
For those workflows, please refer to the repository scripts and the other documentation pages.
Download order
- download dataset
- download pretrained models
- run UNet evaluation
- run GAN evaluation