Closed nfachan closed 2 months ago
Here is another example:
jAug 29 21:28:44.809 ERRO Got error servicing FUSE request. Returning EIO, error: DirectoryDataReader::new at /home/neal/maelstrom/crates/maelstrom-layer-fs/src/dir.rs:32:20
Caused by:
0: open("/home/neal/.cache/maelstrom/worker/artifacts/upper_fs_layer/sha256/b7a11f6ef2f83c22cb1282262fbbe7f070210ba8ea4658c292179fcd3fc23258/65.dir_data.bin")
1: Too many open files (os error 24), args: ["--exact", "--nocapture", "tests::waiting_for_artifacts"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(105) }
Aug 29 21:28:44.809 ERRO Got error servicing FUSE request. Returning EIO, error: DirectoryDataReader::new at /home/neal/maelstrom/crates/maelstrom-layer-fs/src/dir.rs:32:20
Caused by:
0: open("/home/neal/.cache/maelstrom/worker/artifacts/upper_fs_layer/sha256/b7a11f6ef2f83c22cb1282262fbbe7f070210ba8ea4658c292179fcd3fc23258/65.dir_data.bin")
1: Too many open files (os error 24), args: ["--exact", "--nocapture", "tests::running"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(123) }
Aug 29 21:28:44.810 ERRO Got error servicing FUSE request. Returning EIO, error: DirectoryDataReader::new at /home/neal/maelstrom/crates/maelstrom-layer-fs/src/dir.rs:32:20
Caused by:
0: open("/home/neal/.cache/maelstrom/worker/artifacts/upper_fs_layer/sha256/b7a11f6ef2f83c22cb1282262fbbe7f070210ba8ea4658c292179fcd3fc23258/65.dir_data.bin")
1: Too many open files (os error 24), args: ["--exact", "--nocapture", "tests::loop_three_times"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(8) }
Aug 29 21:28:44.810 ERRO Got error servicing FUSE request. Returning EIO, error: DirectoryDataReader::new at /home/neal/maelstrom/crates/maelstrom-layer-fs/src/dir.rs:32:20
Caused by:
0: open("/home/neal/.cache/maelstrom/worker/artifacts/upper_fs_layer/sha256/b7a11f6ef2f83c22cb1282262fbbe7f070210ba8ea4658c292179fcd3fc23258/65.dir_data.bin")
1: Too many open files (os error 24), args: ["--exact", "--nocapture", "tests::expected_count_updates_packages"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(14) }
Aug 29 21:28:44.810 ERRO Got error servicing FUSE request. Returning EIO, error: DirectoryDataReader::new at /home/neal/maelstrom/crates/maelstrom-layer-fs/src/dir.rs:32:20
Caused by:
0: open("/home/neal/.cache/maelstrom/worker/artifacts/upper_fs_layer/sha256/b7a11f6ef2f83c22cb1282262fbbe7f070210ba8ea4658c292179fcd3fc23258/65.dir_data.bin")
1: Too many open files (os error 24), args: ["--exact", "--nocapture", "tests::expected_count_updates_cases"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(55) }
Aug 29 21:28:44.810 ERRO Got error servicing FUSE request. Returning EIO, error: DirectoryDataReader::new at /home/neal/maelstrom/crates/maelstrom-layer-fs/src/dir.rs:32:20
Caused by:
0: open("/home/neal/.cache/maelstrom/worker/artifacts/upper_fs_layer/sha256/b7a11f6ef2f83c22cb1282262fbbe7f070210ba8ea4658c292179fcd3fc23258/65.dir_data.bin")
1: Too many open files (os error 24), args: ["--exact", "--nocapture", "tests::stop_after_1_with_estimate"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(53) }
Aug 29 21:28:44.810 ERRO Got error servicing FUSE request. Returning EIO, error: DirectoryDataReader::new at /home/neal/maelstrom/crates/maelstrom-layer-fs/src/dir.rs:32:20
Caused by:
0: open("/home/neal/.cache/maelstrom/worker/artifacts/upper_fs_layer/sha256/b7a11f6ef2f83c22cb1282262fbbe7f070210ba8ea4658c292179fcd3fc23258/65.dir_data.bin")
1: Too many open files (os error 24), args: ["--exact", "--nocapture", "test_listing::tests::save_of_listing"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(22) }
Aug 29 21:28:44.978 ERRO Failed to get pipe memory, cannot splice, error: Os { code: 24, kind: Uncategorized, message: "Too many open files" }, args: ["--exact", "--nocapture", "metadata::directive::tests::layers_after_image_with_layers"], program: "/maelstrom_test_runner-a3e2391ddb15c556", jid: JobId { cid: ClientId(44), cjid: ClientJobId(16) }
Specific Maelstrom Program?
maelstrom-worker
Bug Description Eventually the worker starts logging errors like this:
It will continue failing every job until it is restarted.
How to Reproduce I don't know how to reproduce it quickly. I just need to use the worker long enough.
Expected Behavior This doesn't happen! :-)
There are all sorts of things it could do: