- Finalized variant workload targets.
- Fix in random_utils helper function.
- For conformer PyTorch Dropout layers set
inplace=True
. - Clear CUDA cache at begining of each trial for PyTorch.
Upgrade CUDA version to CUDA 12.1:
- Upgrade CUDA version in Dockerfiles that will be used for scoring.
- Update Jax and PyTorch package version tags to use local CUDA installation.
Add flag for completely disabling checkpointing.
- Note that we will run with checkpointing off at scoring time.
Update Deepspeech and Conformer variant target setting configurations.
- Note that variant targets are not final.
Fixed bug in scoring code to take best trial in a study for external-tuning ruleset.
Added instructions for submission.
Changed default number of workers for PyTorch data loaders to 0. Running with >0 may lead to incorrect eval results see #732.
Workload variant additions and fixes:
- Add Deepspeech workload variant
- Fix bugs in Imagenet ResNet, WMT and Criteo1tb variants
Add prize qualification logs for external tuning ruleset. Note: FastMRI trials with dropout are not yet added due to #664.
Add missing funcitonality to Docker startup script for self_tuning ruleset. Add self_tuning ruleset option to script that runs all workloads for scoring.
Datasetup fixes.
Fix tests that check training differences in PyTorch and JAX on GPU.
Bug fixes to FastMRI metric calculation and targets.
Added workload variants and targets for ogbg, fastmri, librispeech_conformer, imagenet_resnet, imagenet_vit, criteo1tb to be used as held-out workloads.
First release of the AlgoPerf: Training algorithms benchmarking code.