-
## π Bug
```
python caption.py
WARNING:root:libtpu.so and TPU device found. Setting PJRT_DEVICE=TPU.
Loading checkpoint shards: 100%|ββββββββββββββββββββββββββββββββββββββββββββββββββ| 2/2 [β¦
-
Cmd line:
`torchrun --nproc_per_node 1 example_infilling.py --ckpt_dir CodeLlama-7b-Instruct/ --tokenizer_path CodeLlama-7b-Instruct/tokenizer.model --max_seq_len 512 -- max_batch_size 4`
Error Raβ¦
-
When training with the full dataset, I always encounter one of the following two issues:
1. When training with a single worker, I get an error message
> `(raylet) [2023-04-01 15:58:04,464 E 133β¦
-
There are lots of papers on using oblivious transfer or other mechanisms to generate the private key so that it never lives in one place. An example protocol: https://medium.com/@benny.pinkas/fast-disβ¦
-
# Open Grant Proposal: Delorean Protocol
**Project Name:** Delorean Protocol
**Proposal Category:** Research & protocols
**Individual or Entity Name:** ChainSafe Systems
**Proposer:** willβ¦
-
-
**What happened**:
Tasks which are currently persisting cause annotated tasks to fail.
**What you expected to happen**:
Annotated and non-annotated tasks should run concurrently.
**Minimalβ¦
-
I found that if you spin up a multi-node cluster that uses the AWS Data Protection Provider and all nodes are starting at the exact same time, then you might encounter an issue where more than one "/Mβ¦
-
Hi,
I've tried to run the examples, but I received this error.
```
(CodeLlama) PS C:\Users\marce\OneDrive\mah-docs\CodeLlama\codellama> python -m torch.distributed.run --nproc_per_node 1 example_β¦
-
The smart way would be to implement a yaml serialiser/deserialiser for artemis-odb itself, so as to have component mappings.
But we'll go the easy way first: parse the yaml file and manually check β¦