MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvements in both training algorithms and models.
Traceback (most recent call last):
File "submission_runner.py", line 689, in <module>
app.run(main)
File "/usr/local/lib/python3.8/dist-packages/absl/app.py", line 308, in run
_run_main(main, args)
File "/usr/local/lib/python3.8/dist-packages/absl/app.py", line 254, in _run_main
sys.exit(main(argv))
File "submission_runner.py", line 657, in main
score = score_submission_on_workload(
File "submission_runner.py", line 568, in score_submission_on_workload
timing, metrics = train_once(workload, workload_name,
File "submission_runner.py", line 221, in train_once
model_params, model_state = workload.init_model_fn(
File "/algorithmic-efficiency/algorithmic_efficiency/workloads/criteo1tb/criteo1tb_jax/workload.py", line 102, in init_model_fn
initial_variables = jax.jit(init_fn)(
File "/algorithmic-efficiency/algorithmic_efficiency/workloads/criteo1tb/criteo1tb_jax/models.py", line 91, in __call__
if self.dropout_rate > 0.0 and layer_idx == num_layers_top - 2:
TypeError: '>' not supported between instances of 'NoneType' and 'float'
Steps to Reproduce
Description
Log message:
Looks related to the tuning hparam config.