[Gluster-users] client hung during self heal/replicate
Charles
cshtreck at yahoo.com
Fri Oct 2 18:08:39 UTC 2009
Hi,
- 1 client
- 3 servers
One server had been down when i started writing test data. Once my writes finished I brought up the third server and the self-heal/replication began. However, I noticed that an 'ls -l' from the client hung. Is this normal to have a client hang when 1 of 3 servers is healing ?
my client and server configs at the bottom.
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1859:afr_sh_entry_sync_prepare] replicate: self-healing directory / from subvolume remote2 to 1 other
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod] replicate: creating missing file /test_file2 on remote1
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod] replicate: creating missing file /test_file3 on remote1
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod] replicate: creating missing file /test_file on remote1
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod] replicate: creating missing file /test_file4 on remote1
[2009-10-02 12:29:20] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync] replicate: self-healing metadata of /test_file2 from remote2 to remote1
[2009-10-02 12:29:20] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare] replicate: self-healing file /test_file2 from subvolume remote2 to 1 other
[2009-10-02 12:29:20] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync] replicate: self-healing metadata of /test_file from remote2 to remote1
[2009-10-02 12:29:20] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare] replicate: self-healing file /test_file from subvolume remote2 to 1 other
[2009-10-02 12:29:33] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync] replicate: self-healing metadata of /test_file3 from remote2 to remote1
[2009-10-02 12:29:33] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare] replicate: self-healing file /test_file3 from subvolume remote2 to 1 other
[2009-10-02 12:29:33] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync] replicate: self-healing metadata of /test_file4 from remote2 to remote1
[2009-10-02 12:29:33] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare] replicate: self-healing file /test_file4 from subvolume remote2 to 1 other
# client config
# file: /etc/glusterfs/glusterfs.vol
volume remote1
type protocol/client
option transport-type tcp
option remote-host rh1
option remote-subvolume brick
end-volume
volume remote2
type protocol/client
option transport-type tcp
option remote-host rh2
option remote-subvolume brick
end-volume
volume remote3
type protocol/client
option transport-type tcp
option remote-host rh3
option remote-subvolume brick
end-volume
volume replicate
type cluster/replicate
subvolumes remote1 remote2 remote3
end-volume
volume writebehind
type performance/write-behind
option window-size 1MB
subvolumes replicate
end-volume
volume cache
type performance/io-cache
option cache-size 512MB
subvolumes writebehind
end-volume
#
# server config
#
volume posix
type storage/posix
option directory /bigpartition
end-volume
volume locks
type features/locks
subvolumes posix
end-volume
volume brick
type performance/io-threads
option thread-count 8
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp
option auth.addr.brick.allow *
subvolumes brick
end-volume
Thanks!
=cm
More information about the Gluster-users
mailing list