[Gluster-users] DEBUG log for replicate

eagleeyes eagleeyes at 126.com
Wed Mar 4 07:59:41 UTC 2009


Hello 
    
The directory of GFS server export was on one server .
client ARF 
volume client1
  type protocol/client
  option transport-type tcp     
  option remote-host 172.20.92.249      
  option transport.socket.remote-port 6996              
  option remote-subvolume brick1       
end-volume

volume client2
  type protocol/client
  option transport-type tcp     
  option remote-host 172.20.92.249      
  option transport.socket.remote-port 6997              
  option remote-subvolume brick2        
end-volume
volume client3
  type protocol/client
  option transport-type tcp     
  option remote-host 172.20.92.249      
  option transport.socket.remote-port 6998              
  option remote-subvolume brick3        
end-volume

volume client4
  type protocol/client
  option transport-type tcp     
  option remote-host 172.20.92.249      
  option transport.socket.remote-port 6999              
  option remote-subvolume brick4        
end-volume

volume  ns1 
 type protocol/client
  option transport-type tcp     
  option remote-host 172.20.92.249      
  option transport.socket.remote-port 6996              
  option remote-subvolume name1        
end-volume

volume  ns2 
 type protocol/client
  option transport-type tcp     
  option remote-host 172.20.92.249     
  option transport.socket.remote-port 6997              
  option remote-subvolume name2        
end-volume

volume rep1
  type cluster/replicate
  option  data-self-heal on 
  option  entry-self-heal  on
  option  metadata-self-heal  on
  option  data-lock-server-count 2
  option  entry-lock-server-count 2
  subvolumes client1 client2 
end-volume

volume rep2
  type cluster/replicate
  option  data-self-heal on 
  option  entry-self-heal  on
  option  metadata-self-heal  on
  option  data-lock-server-count 2
  option  entry-lock-server-count 2
  subvolumes client3 client4 
end-volume

volume rep-ns
  type cluster/replicate
  option  data-self-heal on
  option  entry-self-heal  on 
  option  metadata-self-heal  on
  option  data-lock-server-count 2
  option  entry-lock-server-count 2
  subvolumes ns1 ns2 
end-volume
 
olume bricks
  type cluster/unify
  option namespace rep-ns # this will not be storage child of unify.
  subvolumes rep1 rep2
  option self-heal background # foreground off # default is foreground
  option scheduler rr
end-volume 

  When i test  replicate mode , i "rm " a file in GFS server ,and execute "ll -h " in GFS client ,the DEBUG log  is this :

2009-03-04 15:38:00 D [fuse-bridge.c:368:fuse_entry_cbk] glusterfs-fuse: 41: LOOKUP() / => 1 (1)
2009-03-04 15:38:00 D [fuse-bridge.c:1738:fuse_opendir] glusterfs-fuse: 42: OPENDIR /
2009-03-04 15:38:00 D [fuse-bridge.c:652:fuse_fd_cbk] glusterfs-fuse: 42: OPENDIR() / => 0x8280cc0
2009-03-04 15:38:00 D [fuse-bridge.c:368:fuse_entry_cbk] glusterfs-fuse: 43: LOOKUP() / => 1 (1)
2009-03-04 15:38:00 D [fuse-bridge.c:1825:fuse_readdir] glusterfs-fuse: 44: READDIR (0x8280cc0, size=4096, offset=0)
2009-03-04 15:38:00 D [fuse-bridge.c:1771:fuse_readdir_cbk] glusterfs-fuse: 44: READDIR => 6/4096,0
2009-03-04 15:38:00 D [fuse-bridge.c:1825:fuse_readdir] glusterfs-fuse: 45: READDIR (0x8280cc0, size=4096, offset=2147483647)
2009-03-04 15:38:00 D [fuse-bridge.c:1771:fuse_readdir_cbk] glusterfs-fuse: 45: READDIR => 0/4096,2147483647
2009-03-04 15:38:00 D [fuse-bridge.c:1843:fuse_releasedir] glusterfs-fuse: 46: RELEASEDIR 0x8280cc0
2009-03-04 15:38:00 D [inode.c:293:__inode_activate] fuse/inode: activating inode(3538958), lru=3/0 active=2 purge=0
2009-03-04 15:38:00 D [fuse-bridge.c:461:fuse_lookup] glusterfs-fuse: 47: LOOKUP /11(3538958)
2009-03-04 15:38:00 D [afr-self-heal-common.c:1041:afr_self_heal] rep1: performing self heal on /11 (metadata=1 data=1 entry=1)
2009-03-04 15:38:00 D [afr-self-heal-common.c:998:afr_self_heal_missing_entries] rep1: attempting to recreate missing entries for path=/11
2009-03-04 15:38:00 D [afr-self-heal-common.c:962:sh_missing_entries_lk_cbk] rep1: inode of /11 on child 136837152 locked
2009-03-04 15:38:00 D [afr-self-heal-common.c:962:sh_missing_entries_lk_cbk] rep1: inode of /11 on child 136839776 locked
2009-03-04 15:38:00 D [afr-self-heal-common.c:915:sh_missing_entries_lookup] rep1: looking up /11 on subvolume client1
2009-03-04 15:38:00 D [afr-self-heal-common.c:915:sh_missing_entries_lookup] rep1: looking up /11 on subvolume client2
2009-03-04 15:38:00 W [afr-self-heal-common.c:871:sh_missing_entries_lookup_cbk] rep1: path /11 on subvolume client1 => -1 (No such file or directory)
2009-03-04 15:38:00 D [afr-self-heal-common.c:863:sh_missing_entries_lookup_cbk] rep1: path /11 on subvolume client2 is of mode 0100644
2009-03-04 15:38:00 D [afr-self-heal-common.c:608:sh_missing_entries_mknod] rep1: mknod /11 mode 0100644 on 1 subvolumes
2009-03-04 15:38:00 D [afr-self-heal-common.c:555:sh_missing_entries_newentry_cbk] rep1: chown /11 to 0 0 on subvolume client1
2009-03-04 15:38:00 D [afr-self-heal-common.c:502:sh_missing_entries_finish] rep1: unlocking 1/11 on subvolume client1
2009-03-04 15:38:00 D [afr-self-heal-common.c:502:sh_missing_entries_finish] rep1: unlocking 1/11 on subvolume client2
2009-03-04 15:38:00 D [afr-self-heal-common.c:441:afr_sh_missing_entries_done] rep1: proceeding to metadata check on /11
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:752:afr_sh_metadata_lock] rep1: locking /11 on subvolume client1
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:752:afr_sh_metadata_lock] rep1: locking /11 on subvolume client2
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:706:afr_sh_metadata_lk_cbk] rep1: inode of /11 on child 0 locked
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:706:afr_sh_metadata_lk_cbk] rep1: inode of /11 on child 1 locked
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:658:afr_sh_metadata_lookup] rep1: looking up /11 on client1
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:658:afr_sh_metadata_lookup] rep1: looking up /11 on client2
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:604:afr_sh_metadata_lookup_cbk] rep1: path /11 on subvolume client1 is of mode 0100644
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:604:afr_sh_metadata_lookup_cbk] rep1: path /11 on subvolume client2 is of mode 0100644
2009-03-04 15:38:00 D [afr-self-heal-common.c:170:afr_sh_print_pending_matrix] rep1: pending_matrix: [ 0 0 ]
2009-03-04 15:38:00 D [afr-self-heal-common.c:170:afr_sh_print_pending_matrix] rep1: pending_matrix: [ 0 0 ]
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:491:afr_sh_metadata_sync_prepare] rep1: syncing metadata of /11 from subvolume client2 to 1 active sinks
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:383:afr_sh_metadata_sync] rep1: syncing metadata of /11 from client2 to client1
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:249:afr_sh_metadata_erase_pending] rep1: erasing pending flags from /11 on client1
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:249:afr_sh_metadata_erase_pending] rep1: erasing pending flags from /11 on client2
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:156:afr_sh_metadata_finish] rep1: unlocking /11 on subvolume client1
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:156:afr_sh_metadata_finish] rep1: unlocking /11 on subvolume client2
2009-03-04 15:38:00 D [afr-self-heal-metadata.c:83:afr_sh_metadata_done] rep1: proceeding to data check on /11
2009-03-04 15:38:00 D [afr-self-heal-data.c:992:afr_sh_data_lock] rep1: locking /11 on subvolume client1
2009-03-04 15:38:00 D [afr-self-heal-data.c:992:afr_sh_data_lock] rep1: locking /11 on subvolume client2
2009-03-04 15:38:00 D [afr-self-heal-data.c:944:afr_sh_data_lock_cbk] rep1: inode of /11 on child 0 locked
2009-03-04 15:38:00 D [afr-self-heal-data.c:944:afr_sh_data_lock_cbk] rep1: inode of /11 on child 1 locked
2009-03-04 15:38:00 D [afr-self-heal-common.c:170:afr_sh_print_pending_matrix] rep1: pending_matrix: [ 0 0 ]
2009-03-04 15:38:00 D [afr-self-heal-common.c:170:afr_sh_print_pending_matrix] rep1: pending_matrix: [ 0 0 ]
2009-03-04 15:38:00 D [afr-self-heal-data.c:752:afr_sh_data_sync_prepare] rep1: syncing data of /11 from subvolume client2 to 1 active sinks
2009-03-04 15:38:00 D [afr-self-heal-data.c:642:afr_sh_data_open_cbk] rep1: fd for /11 opened, commencing sync
2009-03-04 15:38:00 W [afr-self-heal-data.c:646:afr_sh_data_open_cbk] rep1: sourcing file /11 from client2 to other sinks
2009-03-04 15:38:00 D [afr-self-heal-data.c:501:afr_sh_data_read_cbk] rep1: read 0 bytes of data from /11 on child 1, offset 0
2009-03-04 15:38:00 D [afr-self-heal-data.c:379:afr_sh_data_trim_cbk] rep1: ftruncate of /11 on subvolume client1 completed
2009-03-04 15:38:00 D [afr-self-heal-data.c:328:afr_sh_data_erase_pending] rep1: erasing pending flags from /11 on client1
2009-03-04 15:38:00 D [afr-self-heal-data.c:328:afr_sh_data_erase_pending] rep1: erasing pending flags from /11 on client2
2009-03-04 15:38:00 D [afr-self-heal-data.c:252:afr_sh_data_finish] rep1: finishing data selfheal of /11
2009-03-04 15:38:00 D [afr-self-heal-data.c:228:afr_sh_data_unlock] rep1: unlocking /11 on subvolume client1
2009-03-04 15:38:00 D [afr-self-heal-data.c:228:afr_sh_data_unlock] rep1: unlocking /11 on subvolume client2
2009-03-04 15:38:00 D [afr-self-heal-data.c:185:afr_sh_data_unlck_cbk] rep1: inode of /11 on child 0 locked
2009-03-04 15:38:00 D [afr-self-heal-data.c:185:afr_sh_data_unlck_cbk] rep1: inode of /11 on child 1 locked
2009-03-04 15:38:00 D [afr-self-heal-data.c:134:afr_sh_data_close] rep1: closing fd of /11 on client2
2009-03-04 15:38:00 D [afr-self-heal-data.c:149:afr_sh_data_close] rep1: closing fd of /11 on client1
2009-03-04 15:38:00 D [afr-self-heal-data.c:70:afr_sh_data_done] rep1: self heal of /11 completed
2009-03-04 15:38:00 D [fuse-bridge.c:368:fuse_entry_cbk] glusterfs-fuse: 47: LOOKUP() /11 => 3538958 (3538958)
2009-03-04 15:38:00 D [inode.c:112:__dentry_unhash] fuse/inode: dentry unhashed 11 (3538958)
2009-03-04 15:38:00 D [inode.c:94:__dentry_hash] fuse/inode: dentry hashed 11 (3538958)
2009-03-04 15:38:00 D [inode.c:312:__inode_passivate] fuse/inode: passivating inode(3538958) lru=4/0 active=1 purge=0
2009-03-04 15:38:00 D [inode.c:293:__inode_activate] fuse/inode: activating inode(3538958), lru=3/0 active=2 purge=0
2009-03-04 15:38:00 D [fuse-bridge.c:1512:fuse_open] glusterfs-fuse: 48: OPEN /11
2009-03-04 15:38:00 D [fuse-bridge.c:652:fuse_fd_cbk] glusterfs-fuse: 48: OPEN() /11 => 0x827e918
2009-03-04 15:38:00 D [fuse-bridge.c:1573:fuse_readv] glusterfs-fuse: 49: READ (0x827e918, size=4096, offset=0)
2009-03-04 15:38:00 D [fuse-bridge.c:1538:fuse_readv_cbk] glusterfs-fuse: 49: READ => 0/4096,0/88
2009-03-04 15:38:00 D [fuse-bridge.c:1657:fuse_flush] glusterfs-fuse: 50: FLUSH 0x827e918
2009-03-04 15:38:00 D [fuse-bridge.c:896:fuse_err_cbk] glusterfs-fuse: 50: FLUSH() ERR => 0
2009-03-04 15:38:00 D [fuse-bridge.c:1677:fuse_release] glusterfs-fuse: 51: RELEASE 0x827e918

Why  D [afr-self-heal-data.c:501:afr_sh_data_read_cbk] rep1: read 0 bytes of data from /11 on child 1, offset 0 ???

Wait for your return ,thanks a lot 
2009-03-04 



eagleeyes 
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090304/a35d4158/attachment.html>


More information about the Gluster-users mailing list