Description of problem:
Encountering recurring NFS share freezing due to the error message fsal_fd_global_counter is negative, leading to prolonged production downtime.
Mandatory info:- The output of the gluster volume info command:
Volume Name: swarm-utils
Type: Replicate
Volume ID: 59294449-a062-4804-b530-87cc3c7bb378
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: swarm-fs02.srv.dhw.de:/swarm/utils
Brick2: swarm-fs03.srv.dhw.de:/swarm/utils
Brick3: swarm-fs01.srv.dhw.de:/swarm/utils
Options Reconfigured:
storage.owner-uid: 65534
transport.address-family: inet
storage.fips-mode-rchecksum: on
nfs.disable: on
performance.client-io-threads: off
Volume Name: swarm-volumes
Type: Replicate
Volume ID: 3671e28c-0cde-44ed-86f0-f41ffa793051
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: swarm-fs02.srv.dhw.de:/swarm/volumes
Brick2: swarm-fs03.srv.dhw.de:/swarm/volumes
Brick3: swarm-fs01.srv.dhw.de:/swarm/volumes
Options Reconfigured:
storage.owner-uid: 65534
transport.address-family: inet
storage.fips-mode-rchecksum: on
nfs.disable: on
performance.client-io-threads: off
- The output of the gluster volume status command:
Status of volume: swarm-utils
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick swarm-fs02.srv.dhw.de:/swarm/utils 58650 0 Y 733
Brick swarm-fs03.srv.dhw.de:/swarm/utils 50427 0 Y 733
Brick swarm-fs01.srv.dhw.de:/swarm/utils 55932 0 Y 792
Self-heal Daemon on localhost N/A N/A Y 826
Self-heal Daemon on 10.15.29.13 N/A N/A Y 803
Self-heal Daemon on 10.15.29.12 N/A N/A Y 5987
Task Status of Volume swarm-utils
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: swarm-volumes
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick swarm-fs02.srv.dhw.de:/swarm/volumes 52760 0 Y 766
Brick swarm-fs03.srv.dhw.de:/swarm/volumes 52048 0 Y 787
Brick swarm-fs01.srv.dhw.de:/swarm/volumes 53120 0 Y 807
Self-heal Daemon on localhost N/A N/A Y 826
Self-heal Daemon on 10.15.29.13 N/A N/A Y 803
Self-heal Daemon on 10.15.29.12 N/A N/A Y 5987
Task Status of Volume swarm-volumes
------------------------------------------------------------------------------
There are no active volume tasks
**- Provide logs present on following locations of client and server nodes -
/var/log/glusterfs/glusterd.log
The message "I [MSGID: 106496] [glusterd-handshake.c:954:__server_getspec] 0-management: Received mount request for volume shd/swarm-utils" repeated 5 times between [2023-11-21 23:00:26.893764 +0000] and [2023-11-21 23:00:27.104694 +0000]
The message "I [MSGID: 106496] [glusterd-handshake.c:954:__server_getspec] 0-management: Received mount request for volume shd/swarm-volumes" repeated 5 times between [2023-11-21 23:00:26.894494 +0000] and [2023-11-21 23:00:27.104767 +0000]
[2023-11-22 07:00:27.057181 +0000] I [MSGID: 106061] [glusterd-utils.c:10724:glusterd_volume_status_copy_to_op_ctx_dict] 0-management: Dict get failed [{Key=count}]
[2023-11-22 07:00:27.057639 +0000] I [MSGID: 106499] [glusterd-handler.c:4372:__glusterd_handle_status_volume] 0-management: Received status volume req for volume swarm-utils
[2023-11-22 07:00:27.060479 +0000] I [MSGID: 106499] [glusterd-handler.c:4372:__glusterd_handle_status_volume] 0-management: Received status volume req for volume swarm-volumes
**- Is there any crash ? Provide the backtrace and coredump
/var/log/ganesha/ganesha.log
Description of problem: Encountering recurring NFS share freezing due to the error message
fsal_fd_global_counter is negative
, leading to prolonged production downtime.Mandatory info: - The output of the
gluster volume info
command:- The output of the
gluster volume status
command:**- Provide logs present on following locations of client and server nodes - /var/log/glusterfs/glusterd.log
**- Is there any crash ? Provide the backtrace and coredump /var/log/ganesha/ganesha.log
Additional info:
- The operating system / glusterfs version: Ubuntu 20.04 LTS Gluster 10.5 Ganesha 5.7