-
Traceback (most recent call last):
File "train.py", line 513, in
train(hyp, opt, device, tb_writer, wandb)
File "train.py", line 299, in train
scaler.step(optimizer) # optimizer.step…
-
Hi @gavinfay,
The simulation loop for the sdmTMB change scenarios has been added through commit [278197b](https://github.com/thefaylab/sseep-analysis/commit/278197b3022fe5d22242aafabe53cc05b9272f1…
-
Currently, I am trying to use `FMUParameterRegistrator` to do parameter calibration/estimation but encounter some issues.
I do experiments on both ME-type & CS-type FMU:
- **For ME-type FMU**
…
-
A priority for future development should be to make this package compatible with arbitrary derivative-free `Optim.jl` or `BlackBoxOptim.jl` optimization algorithms. These problems are formulated not a…
-
Hello,
I want to add l2 regularisation .Can you tell me where can I add this line:
optimizer = torch.optim.Adam(model.parameters(), lr=1e-4, weight_decay=1e-5)
-
**Describe the bug**
I have trained a 1.3B model on 64 A100 80G Gpus, I export the saved checkpoints except the deepspeed zero-optimal states, the exported ckpts structure is same as your opensource …
-
### 🐛 Describe the bug
When a module has a parameter which is a tensor of size 1 and you try to save its FSDP with torch.distributed.checkpoint, you get the following exception:
```
NotImplemente…
-
根据官方教程进行复现 进行微调时出现grad_norm:nan
参数配置如下:
# Model
pretrained_model_name_or_path = 'internlm/internlm2-chat-7b'
use_varlen_attn = False
# Data
data_path = 'data'
prompt_template = PROMPT_TEMPLAT…
-
Dear:
when I run this code, got the following error,
site-packages/torch/optim/adam.py", line 86, in step
exp_avg, exp_avg_sq = state['exp_avg'], state['exp_avg_sq']
KeyError: 'exp_avg'
w…
-
I wanted to continue training the model with the saved optimizer, but it crashed. The traceback is shown as follows:
Traceback (most recent call last):
File "lgesql/text2sql.py", line 105, in
…