autonomys / subspace

Subspace Network reference implementation
https://subspace.network
369 stars 242 forks source link

Farm is likely already in use, make sure no other farmer is using it: Bad file descriptor (os error 9) #2582

Closed Tocard closed 6 months ago

Tocard commented 6 months ago

Hello there :)

I do have a strange behavior, dunno if this come from subspace itself or from my implementation. In doubt I'm posting is.

Running into proxmox => Ubuntu server 22. using nfsV4 (thanks to Qwiin) februad19 release & cli mode


[Unit]
Description="subspace farmer"
After=network-online.target

[Service]
User=chimera
Group=chimera
WorkingDirectory=/opt/chimera/farmer/
ExecStart=subspace_farmer farm \
    --node-rpc-url ws://127.0.0.1:9944 \
     path=/mnt/subspace_001,size=1.8T \
     path=/mnt/subspace_002,size=1.8T \
     path=/mnt/subspace_003,size=900G \
     path=/mnt/hydras_alpha_fikwot4T_baie_top_left/alpha,size=3.7T \
     path=/opt/chimera/farmer/ssd,size=800G \
     path=/opt/chimera/farmer/nvme,size=900G \
     --reward-address st8VCipcz7xezUnM73T7szCiyNau51YYJAGovKGETUjuPy5oj \
    --listen-on /ip4/0.0.0.0/tcp/30533 \
    --farm-during-initial-plotting false \
    --prometheus-listen-on 0.0.0.0:9081

StandardOutput=append:/var/log/chimera/subspace_farmer/farmer.log
StandardError=append:/var/log/chimera/subspace_farmer/farmer.log
LimitNOFILE=infinity

Restart=on-failure
RestartSec=5s

[Install]
WantedBy=multi-user.target

The issue is this, but i do not have any farm / plot or whatever, so...


Error: Farm is likely already in use, make sure no other farmer is using it: Bad file descriptor (os error 9)
2024-03-03T15:31:08.783640Z  INFO subspace_farmer::commands::farm: Connecting to node RPC url=ws://127.0.0.1:9944
2024-03-03T15:31:08.789310Z  INFO subspace_networking::constructor: DSN instance configured. allow_non_global_addresses_in_dht=false peer_id=12D3KooWFzL7RdsCZw8mACQdeySk1WF6TzVdvQMKHb7tgtnWtsRS protocol_version=/subspace/2/0c121c75f4ef450f40619e1fca9d1e8e7fbabc42c895bc4790801e85d5a91c34
2024-03-03T15:31:08.789814Z  INFO libp2p_swarm: local_peer_id=12D3KooWFzL7RdsCZw8mACQdeySk1WF6TzVdvQMKHb7tgtnWtsRS
2024-03-03T15:31:08.789968Z  INFO subspace_metrics: Metrics server started. endpoints=[0.0.0.0:9081]
2024-03-03T15:31:08.789997Z  INFO actix_server::builder: starting 2 workers
2024-03-03T15:31:08.790037Z  INFO actix_server::server: Tokio runtime found; starting in existing Tokio runtime
2024-03-03T15:31:09.210305Z  INFO subspace_farmer::commands::farm: Multiple L3 cache groups detected l3_cache_groups=2
Error: Farm is likely already in use, make sure no other farmer is using it: Bad file descriptor (os error 9)
2024-03-03T15:31:14.283171Z  INFO subspace_farmer::commands::farm: Connecting to node RPC url=ws://127.0.0.1:9944
2024-03-03T15:31:14.285862Z  INFO subspace_networking::constructor: DSN instance configured. allow_non_global_addresses_in_dht=false peer_id=12D3KooWFzL7RdsCZw8mACQdeySk1WF6TzVdvQMKHb7tgtnWtsRS protocol_version=/subspace/2/0c121c75f4ef450f40619e1fca9d1e8e7fbabc42c895bc4790801e85d5a91c34
2024-03-03T15:31:14.286303Z  INFO libp2p_swarm: local_peer_id=12D3KooWFzL7RdsCZw8mACQdeySk1WF6TzVdvQMKHb7tgtnWtsRS
2024-03-03T15:31:14.286439Z  INFO subspace_metrics: Metrics server started. endpoints=[0.0.0.0:9081]
2024-03-03T15:31:14.286457Z  INFO actix_server::builder: starting 2 workers
2024-03-03T15:31:14.286488Z  INFO actix_server::server: Tokio runtime found; starting in existing Tokio runtime
2024-03-03T15:31:14.700981Z  INFO subspace_farmer::commands::farm: Multiple L3 cache groups detected l3_cache_groups=2
Error: Farm is likely already in use, make sure no other farmer is using it: Bad file descriptor (os error 9)
nazar-pc commented 6 months ago

If nothing is opening then there is clearly something on your end and NFS was never a supported configuration, so at your risk. As for locking, it is known to be problematic on NFS and for this reason --disable-farm-locking was added, but you need to understand the dangers of it.

Either way this is not an issue with Subspace software and you should create a forum post about this instead of GitHub issue.