[Gluster-users] WG: Strange issu concerning glusterfs 3.5.1 on centos 6.5
Daniel Müller
mueller at tropenklinik.de
Thu Jul 31 11:07:26 UTC 2014
So,
[root at centclust1 ~]# ifconfig
eth0 Link encap:Ethernet Hardware Adresse 00:25:90:80:D9:E8
inet Adresse:172.17.2.30 Bcast:172.17.2.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:d9e8/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:3506528 errors:0 dropped:0 overruns:0 frame:0
TX packets:169905 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:476128477 (454.0 MiB) TX bytes:18788266 (17.9 MiB)
Speicher:fe860000-fe880000
eth1 Link encap:Ethernet Hardware Adresse 00:25:90:80:D9:E9
inet Adresse:192.168.135.36 Bcast:192.168.135.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:d9e9/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:381664693 errors:0 dropped:0 overruns:0 frame:0
TX packets:380924973 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:477454156923 (444.6 GiB) TX bytes:476729269342 (443.9 GiB)
Speicher:fe8e0000-fe900000
lo Link encap:Lokale Schleife
inet Adresse:127.0.0.1 Maske:255.0.0.0
inet6 Adresse: ::1/128 Gültigkeitsbereich:Maschine
UP LOOPBACK RUNNING MTU:16436 Metric:1
RX packets:93922879 errors:0 dropped:0 overruns:0 frame:0
TX packets:93922879 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:0
RX bytes:462579764180 (430.8 GiB) TX bytes:462579764180 (430.8 GiB)
[root at centclust2 ~]# ifconfig
eth0 Link encap:Ethernet Hardware Adresse 00:25:90:80:EF:00
inet Adresse:172.17.2.31 Bcast:172.17.2.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:ef00/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:1383117 errors:0 dropped:0 overruns:0 frame:0
TX packets:45828 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:185634714 (177.0 MiB) TX bytes:5357926 (5.1 MiB)
Speicher:fe860000-fe880000
eth1 Link encap:Ethernet Hardware Adresse 00:25:90:80:EF:01
inet Adresse:192.168.135.46 Bcast:192.168.135.255 Maske:255.255.255.0
inet6 Adresse: fe80::225:90ff:fe80:ef01/64 Gültigkeitsbereich:Verbindung
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:340364283 errors:0 dropped:0 overruns:0 frame:0
TX packets:59930672 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:1000
RX bytes:473823738544 (441.2 GiB) TX bytes:9973035418 (9.2 GiB)
Speicher:fe8e0000-fe900000
lo Link encap:Lokale Schleife
inet Adresse:127.0.0.1 Maske:255.0.0.0
inet6 Adresse: ::1/128 Gültigkeitsbereich:Maschine
UP LOOPBACK RUNNING MTU:16436 Metric:1
RX packets:1102979 errors:0 dropped:0 overruns:0 frame:0
TX packets:1102979 errors:0 dropped:0 overruns:0 carrier:0
Kollisionen:0 Sendewarteschlangenlänge:0
RX bytes:126066547 (120.2 MiB) TX bytes:126066547 (120.2 MiB)
[root at centclust1 ~]# route
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 * 255.255.255.0 U 1 0 0 eth1
172.17.2.0 * 255.255.255.0 U 1 0 0 eth0
default s4master 0.0.0.0 UG 0 0 0 eth1
[root at centclust2 ~]# route
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 * 255.255.255.0 U 0 0 0 eth1
172.17.2.0 * 255.255.255.0 U 0 0 0 eth0
link-local * 255.255.0.0 U 1002 0 0 eth0
link-local * 255.255.0.0 U 1003 0 0 eth1
default s4master 0.0.0.0 UG 0 0 0 eth1
[root at centclust1 ~]# route -n
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 0.0.0.0 255.255.255.0 U 1 0 0 eth1
172.17.2.0 0.0.0.0 255.255.255.0 U 1 0 0 eth0
0.0.0.0 192.168.135.230 0.0.0.0 UG 0 0 0 eth1
[root at centclust2 ~]# route -n
Kernel IP Routentabelle
Ziel Router Genmask Flags Metric Ref Use Iface
192.168.135.0 0.0.0.0 255.255.255.0 U 0 0 0 eth1
172.17.2.0 0.0.0.0 255.255.255.0 U 0 0 0 eth0
169.254.0.0 0.0.0.0 255.255.0.0 U 1002 0 0 eth0
169.254.0.0 0.0.0.0 255.255.0.0 U 1003 0 0 eth1
0.0.0.0 192.168.135.230 0.0.0.0 UG 0 0 0 eth1
EDV Daniel Müller
Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus
Paul-Lechler-Str. 24
72076 Tübingen
Tel.: 07071/206-463, Fax: 07071/206-499
eMail: mueller at tropenklinik.de
Internet: www.tropenklinik.de
-----Ursprüngliche Nachricht-----
Von: Krishnan Parthasarathi [mailto:kparthas at redhat.com]
Gesendet: Donnerstag, 31. Juli 2014 12:55
An: mueller at tropenklinik.de
Cc: gluster-devel-bounces at gluster.org; gluster-users at gluster.org
Betreff: Re: [Gluster-users] WG: Strange issu concerning glusterfs 3.5.1 on centos 6.5
Daniel,
Could you provide the following details from your original two NIC setup probed using hostname?
1) output of ifconfig of the two NICs on both the nodes .
2) output of route from both the nodes.
~KP
----- Original Message -----
> Hello and thank you so far,
> What I have recognized is, having more than one nic running this is
> confusing glusterfs 3.5. I never saw this on my glusterfs 3.4 and 3.2
> systems still working.
> So I set up just clean erased gluster with yum glusterfs* erase and did:
> Logged in to my both nodes in the 135 subnet,ex:
> Ssh 192.168.135.36 (centclust1) (172.17.2.30 is the 2nd nic) Ssh
> 192.168.135.46 (centclust2) (172.17.2.31 is the 2nd nic) Started
> gluster on both nodes , service glusterd start.
> Did the peer probe on 192.168.135.36/centclust1:
> Gluster peer probe 192.168.135.46 //Former I did gluster peer probe
> centclust2
> This result in:
> [root at centclust1 ~]# gluster peer status Number of Peers: 1
>
> Hostname: 192.168.135.46
> Uuid: c395c15d-5187-4e5b-b680-57afcb88b881
> State: Peer in Cluster (Connected)
>
> [root at centclust2 backup]# gluster peer status Number of Peers: 1
>
> Hostname: 192.168.135.36
> Uuid: 94d5903b-ebe9-40d6-93bf-c2f2e92909a0
> State: Peer in Cluster (Connected)
> The signifent difference gluster now shows the ip of both nodes
>
> Now I did the create the replicating vol:
> gluster volume create smbcluster replica 2 transport tcp
> 192.168.135.36:/sbu/glusterfs/export
> 192.168.135.46:/sbu/glusterfs/export
> started the volume
> gluster volume status
> Status of volume: smbcluster
> Gluster process Port Online Pid
> ------------------------------------------------------------------------------
> Brick 192.168.135.36:/sbu/glusterfs/export 49152 Y 27421
> Brick 192.168.135.46:/sbu/glusterfs/export 49152 Y 12186
> NFS Server on localhost 2049 Y 27435
> Self-heal Daemon on localhost N/A Y 27439
> NFS Server on 192.168.135.46 2049 Y 12200
> Self-heal Daemon on 192.168.135.46 N/A Y 12204
>
> Task Status of Volume smbcluster
> ----------------------------------------------------------------------
> --------
> There are no active volume tasks
>
> Mounted the volumes:
>
> Centclust1:mount -t glusterfs 192.168.135.36:/smbcluster /mntgluster
> -o acl Centclust2:mount -t glusterfs 192.168.135.46:/smbcluster
> /mntgluster -o acl
>
> And BINGO up and running!!!!!!!
>
>
> EDV Daniel Müller
>
> Leitung EDV
> Tropenklinik Paul-Lechler-Krankenhaus
> Paul-Lechler-Str. 24
> 72076 Tübingen
> Tel.: 07071/206-463, Fax: 07071/206-499
> eMail: mueller at tropenklinik.de
> Internet: www.tropenklinik.de
>
>
>
>
> -----Ursprüngliche Nachricht-----
> Von: Krishnan Parthasarathi [mailto:kparthas at redhat.com]
> Gesendet: Mittwoch, 30. Juli 2014 16:52
> An: mueller at tropenklinik.de
> Cc: gluster-devel-bounces at gluster.org; gluster-users at gluster.org
> Betreff: Re: [Gluster-users] WG: Strange issu concerning glusterfs
> 3.5.1 on centos 6.5
>
> Daniel,
>
> I didn't get a chance to follow up with debugging this issue. I will
> look into this and get back to you. I suspect that there is something
> different about the network layer behaviour in your setup.
>
> ~KP
>
> ----- Original Message -----
> > Just another other test:
> > [root at centclust1 sicherung]# getfattr -d -e hex -m . /sicherung/bu
> > getfattr: Entferne führenden '/' von absoluten Pfadnamen # file:
> > sicherung/bu
> > security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a66696
> > c6
> > 55f743a733000
> > trusted.afr.smbbackup-client-0=0x000000000000000000000000
> > trusted.afr.smbbackup-client-1=0x000000000000000200000001
> > trusted.gfid=0x00000000000000000000000000000001
> > trusted.glusterfs.dht=0x000000010000000000000000ffffffff
> > trusted.glusterfs.volume-id=0x6f51d002e634437db58d9b952693f1df
> >
> > [root at centclust2 glusterfs]# getfattr -d -e hex -m . /sicherung/bu
> > getfattr: Entferne führenden '/' von absoluten Pfadnamen # file:
> > sicherung/bu
> > security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a66696
> > c6
> > 55f743a733000
> > trusted.afr.smbbackup-client-0=0x000000000000000200000001
> > trusted.afr.smbbackup-client-1=0x000000000000000000000000
> > trusted.gfid=0x00000000000000000000000000000001
> > trusted.glusterfs.dht=0x000000010000000000000000ffffffff
> > trusted.glusterfs.volume-id=0x6f51d002e634437db58d9b952693f1df
> >
> > Is this ok?
> >
> > After long testing and doing a /etc/init.d/network restart the
> > replication started once/a short time then ended up!?
> > Any idea???????
> >
> >
> > EDV Daniel Müller
> >
> > Leitung EDV
> > Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
> > 72076 Tübingen
> > Tel.: 07071/206-463, Fax: 07071/206-499
> > eMail: mueller at tropenklinik.de
> > Internet: www.tropenklinik.de
> >
> > "Der Mensch ist die Medizin des Menschen"
> >
> >
> >
> >
> > -----Ursprüngliche Nachricht-----
> > Von: Krishnan Parthasarathi [mailto:kparthas at redhat.com]
> > Gesendet: Mittwoch, 30. Juli 2014 11:09
> > An: mueller at tropenklinik.de
> > Cc: gluster-devel-bounces at gluster.org; gluster-users at gluster.org
> > Betreff: Re: [Gluster-users] WG: Strange issu concerning glusterfs
> > 3.5.1 on centos 6.5
> >
> > Could you provide the output of the following command?
> >
> > netstat -ntap | grep gluster
> >
> > This should tell us if glusterfsd processes (bricks) are listening
> > on all interfaces.
> >
> > ~KP
> >
> > ----- Original Message -----
> > > Just one idea
> > > I add a second NIC with a 172.2.17... adress on both machines.
> > > Could this cause the trouble!?
> > >
> > > EDV Daniel Müller
> > >
> > > Leitung EDV
> > > Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
> > > 72076 Tübingen
> > > Tel.: 07071/206-463, Fax: 07071/206-499
> > > eMail: mueller at tropenklinik.de
> > > Internet: www.tropenklinik.de
> > >
> > >
> > >
> > >
> > > -----Ursprüngliche Nachricht-----
> > > Von: Krishnan Parthasarathi [mailto:kparthas at redhat.com]
> > > Gesendet: Mittwoch, 30. Juli 2014 09:29
> > > An: mueller at tropenklinik.de
> > > Cc: gluster-devel-bounces at gluster.org; gluster-users at gluster.org
> > > Betreff: Re: [Gluster-users] WG: Strange issu concerning glusterfs
> > > 3.5.1 on centos 6.5
> > >
> > > Daniel,
> > >
> > > From a quick look, I see that glustershd and the nfs client is
> > > unable to connect to one of the bricks. This is resulting in data
> > > from mounts being written to local bricks only.
> > > I should have asked this before, could you provide the bricks logs
> > > as well?
> > >
> > > Could you also try to connect to the bricks using telnet?
> > > For eg, from centclust1, telnet centclust2 <brick-port>.
> > >
> > > ~KP
> > >
> > > ----- Original Message -----
> > > > So my logs. I disable ssl meanwhile but it is the same situation.
> > > > No replication!?
> > > >
> > > >
> > > >
> > > > EDV Daniel Müller
> > > >
> > > > Leitung EDV
> > > > Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
> > > > 72076 Tübingen
> > > > Tel.: 07071/206-463, Fax: 07071/206-499
> > > > eMail: mueller at tropenklinik.de
> > > > Internet: www.tropenklinik.de
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > -----Ursprüngliche Nachricht-----
> > > > Von: Krishnan Parthasarathi [mailto:kparthas at redhat.com]
> > > > Gesendet: Mittwoch, 30. Juli 2014 08:56
> > > > An: mueller at tropenklinik.de
> > > > Cc: gluster-users at gluster.org; gluster-devel-bounces at gluster.org
> > > > Betreff: Re: [Gluster-users] WG: Strange issu concerning
> > > > glusterfs
> > > > 3.5.1 on centos 6.5
> > > >
> > > > Could you attach the entire mount and glustershd log files to
> > > > this thread?
> > > >
> > > > ~KP
> > > >
> > > > ----- Original Message -----
> > > > > NO ONE!??
> > > > > This is an entry of my glustershd.log:
> > > > > [2014-07-30 06:40:59.294334] W
> > > > > [client-handshake.c:1846:client_dump_version_cbk]
> > > > > 0-smbbackup-client-1:
> > > > > received RPC status error
> > > > > [2014-07-30 06:40:59.294352] I
> > > > > [client.c:2229:client_rpc_notify]
> > > > > 0-smbbackup-client-1: disconnected from 172.17.2.31:49152.
> > > > > Client process will keep trying to connect to glusterd until
> > > > > brick's port is available
> > > > >
> > > > >
> > > > > This is from mnt-sicherung.log:
> > > > > [2014-07-30 06:40:38.259850] E [socket.c:2820:socket_connect]
> > > > > 1-smbbackup-client-0: connection attempt on 172.17.2.30:24007
> > > > > failed, (Connection timed out) [2014-07-30 06:40:41.275120] I
> > > > > [rpc-clnt.c:1729:rpc_clnt_reconfig]
> > > > > 1-smbbackup-client-0: changing port to 49152 (from 0)
> > > > >
> > > > > [root at centclust1 sicherung]# gluster --remote-host=centclust1
> > > > > peer status Number of Peers: 1
> > > > >
> > > > > Hostname: centclust2
> > > > > Uuid: 4f15e9bd-9b5a-435b-83d2-4ed202c66b11
> > > > > State: Peer in Cluster (Connected)
> > > > >
> > > > > [root at centclust1 sicherung]# gluster --remote-host=centclust2
> > > > > peer status Number of Peers: 1
> > > > >
> > > > > Hostname: 172.17.2.30
> > > > > Uuid: 99fe6a2c-df7e-4475-a7bc-a35abba620fb
> > > > > State: Peer in Cluster (Connected)
> > > > >
> > > > > [root at centclust1 ssl]# ps aux | grep gluster
> > > > > root 13655 0.0 0.0 413848 16872 ? Ssl 08:10 0:00
> > > > > /usr/sbin/glusterd --pid-file=/var/run/glusterd.pid
> > > > > root 13958 0.0 0.0 12139920 44812 ? Ssl 08:11 0:00
> > > > > /usr/sbin/glusterfsd -s centclust1.tplk.loc --volfile-id
> > > > > smbbackup.centclust1.tplk.loc.sicherung-bu -p
> > > > > /var/lib/glusterd/vols/smbbackup/run/centclust1.tplk.loc-sicherung-bu.
> > > > > pid -S /var/run/4c65260e12e2d3a9a5549446f491f383.socket
> > > > > --brick-name /sicherung/bu -l
> > > > > /var/log/glusterfs/bricks/sicherung-bu.log
> > > > > --xlator-option
> > > > > *-posix.glusterd-uuid=99fe6a2c-df7e-4475-a7bc-a35abba620fb
> > > > > --brick-port
> > > > > 49152 --xlator-option smbbackup-server.listen-port=49152
> > > > > root 13972 0.0 0.0 815748 58252 ? Ssl 08:11 0:00
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
> > > > > /var/lib/glusterd/nfs/run/nfs.pid -l
> > > > > /var/log/glusterfs/nfs.log -S /var/run/ee6f37fc79b9cb1968eca387930b39fb.socket
> > > > > root 13976 0.0 0.0 831160 29492 ? Ssl 08:11 0:00
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id
> > > > > gluster/glustershd -p
> > > > > /var/lib/glusterd/glustershd/run/glustershd.pid -l
> > > > > /var/log/glusterfs/glustershd.log -S
> > > > > /var/run/aa970d146eb23ba7124e6c4511879850.socket --xlator-option *replicate*.node-uuid=99fe6a2c-df7e-4475-a7bc-a35abba620fb
> > > > > root 15781 0.0 0.0 105308 932 pts/1 S+ 08:47 0:00 grep
> > > > > gluster
> > > > > root 29283 0.0 0.0 451116 56812 ? Ssl Jul29 0:21
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
> > > > > /var/lib/glusterd/nfs/run/nfs.pid -l
> > > > > /var/log/glusterfs/nfs.log -S /var/run/a7fcb1d1d3a769d28df80b85ae5d13c4.socket
> > > > > root 29287 0.0 0.0 335432 25848 ? Ssl Jul29 0:21
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id
> > > > > gluster/glustershd -p
> > > > > /var/lib/glusterd/glustershd/run/glustershd.pid -l
> > > > > /var/log/glusterfs/glustershd.log -S
> > > > > /var/run/833e60f976365c2a307f92fb233942a2.socket --xlator-option *replicate*.node-uuid=64b1a7eb-2df3-47bd-9379-39c29e5a001a
> > > > > root 31698 0.0 0.0 1438392 57952 ? Ssl Jul29 0:12
> > > > > /usr/sbin/glusterfs --acl --volfile-server=centclust1.tplk.loc
> > > > > --volfile-id=/smbbackup /mnt/sicherung
> > > > >
> > > > > [root at centclust2 glusterfs]# ps aux | grep gluster
> > > > > root 1561 0.0 0.0 1481492 60152 ? Ssl Jul29 0:12
> > > > > /usr/sbin/glusterfs --acl --volfile-server=centclust2.tplk.loc
> > > > > --volfile-id=/smbbackup /mnt/sicherung
> > > > > root 15656 0.0 0.0 413848 16832 ? Ssl 08:11 0:01
> > > > > /usr/sbin/glusterd --pid-file=/var/run/glusterd.pid
> > > > > root 15942 0.0 0.0 12508704 43860 ? Ssl 08:11 0:00
> > > > > /usr/sbin/glusterfsd -s centclust2.tplk.loc --volfile-id
> > > > > smbbackup.centclust2.tplk.loc.sicherung-bu -p
> > > > > /var/lib/glusterd/vols/smbbackup/run/centclust2.tplk.loc-sicherung-bu.
> > > > > pid -S /var/run/40a554af3860eddd5794b524576d0520.socket
> > > > > --brick-name /sicherung/bu -l
> > > > > /var/log/glusterfs/bricks/sicherung-bu.log
> > > > > --xlator-option
> > > > > *-posix.glusterd-uuid=4f15e9bd-9b5a-435b-83d2-4ed202c66b11
> > > > > --brick-port
> > > > > 49152 --xlator-option smbbackup-server.listen-port=49152
> > > > > root 15956 0.0 0.0 825992 57496 ? Ssl 08:11 0:00
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
> > > > > /var/lib/glusterd/nfs/run/nfs.pid -l
> > > > > /var/log/glusterfs/nfs.log -S /var/run/602d1d8ba7b80ded2b70305ed7417cf5.socket
> > > > > root 15960 0.0 0.0 841404 26760 ? Ssl 08:11 0:00
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id
> > > > > gluster/glustershd -p
> > > > > /var/lib/glusterd/glustershd/run/glustershd.pid -l
> > > > > /var/log/glusterfs/glustershd.log -S
> > > > > /var/run/504d01c7f7df8b8306951cc2aaeaf52c.socket
> > > > > --xlator-option
> > > > > *replicate*.node-uuid=4f15e9bd-9b5a-435b-83d2-4ed202c66b11
> > > > > root 17728 0.0 0.0 105312 936 pts/0 S+ 08:48 0:00 grep
> > > > > gluster
> > > > > root 32363 0.0 0.0 451100 55584 ? Ssl Jul29 0:21
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p
> > > > > /var/lib/glusterd/nfs/run/nfs.pid -l
> > > > > /var/log/glusterfs/nfs.log -S /var/run/73054288d1cadfb87b4b9827bd205c7b.socket
> > > > > root 32370 0.0 0.0 335432 26220 ? Ssl Jul29 0:21
> > > > > /usr/sbin/glusterfs -s localhost --volfile-id
> > > > > gluster/glustershd -p
> > > > > /var/lib/glusterd/glustershd/run/glustershd.pid -l
> > > > > /var/log/glusterfs/glustershd.log -S
> > > > > /var/run/de1427ce373c792c76c38b12c106f029.socket
> > > > > --xlator-option
> > > > > *replicate*.node-uuid=83e6d78c-0119-4537-8922-b3e731718864
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > Leitung EDV
> > > > > Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
> > > > > 72076 Tübingen
> > > > > Tel.: 07071/206-463, Fax: 07071/206-499
> > > > > eMail: mueller at tropenklinik.de
> > > > > Internet: www.tropenklinik.de
> > > > >
> > > > >
> > > > >
> > > > > -----Ursprüngliche Nachricht-----
> > > > > Von: Daniel Müller [mailto:mueller at tropenklinik.de]
> > > > > Gesendet: Dienstag, 29. Juli 2014 16:02
> > > > > An: 'gluster-users at gluster.org'
> > > > > Betreff: Strange issu concerning glusterfs 3.5.1 on centos 6.5
> > > > >
> > > > > Dear all,
> > > > >
> > > > > there is a strange issue centos6.5 and glusterfs 3.5.1:
> > > > >
> > > > > glusterd -V
> > > > > glusterfs 3.5.1 built on Jun 24 2014 15:09:41 Repository revision:
> > > > > git://git.gluster.com/glusterfs.git
> > > > > Copyright (c) 2006-2013 Red Hat, Inc. <http://www.redhat.com/>
> > > > > GlusterFS comes with ABSOLUTELY NO WARRANTY.
> > > > > It is licensed to you under your choice of the GNU Lesser
> > > > > General Public License, version 3 or any later version (LGPLv3
> > > > > or later), or the GNU General Public License, version 2
> > > > > (GPLv2), in all cases as published by the Free Software
> > > > > Foundation
> > > > >
> > > > > I try to set up a replicated 2 brick vol on two centos 6.5 server.
> > > > > I can probe well and my nodes are reporting no errors:
> > > > >
> > > > > [root at centclust1 mnt]# gluster peer status Number of Peers: 1
> > > > >
> > > > > Hostname: centclust2
> > > > > Uuid: 4f15e9bd-9b5a-435b-83d2-4ed202c66b11
> > > > > State: Peer in Cluster (Connected)
> > > > >
> > > > > [root at centclust2 sicherung]# gluster peer status Number of Peers:
> > > > > 1
> > > > >
> > > > > Hostname: 172.17.2.30
> > > > > Uuid: 99fe6a2c-df7e-4475-a7bc-a35abba620fb
> > > > > State: Peer in Cluster (Connected)
> > > > >
> > > > > Now I set up a replicating VOl on an XFS-Disk: /dev/sdb1 on
> > > > > /sicherung type xfs (rw)
> > > > >
> > > > > gluster volume create smbbackup replica 2 transport tcp
> > > > > centclust1.tplk.loc:/sicherung/bu
> > > > > centclust2.tplk.loc:/sicherung/bu
> > > > >
> > > > > gluster volume smbbackup status reports ok:
> > > > >
> > > > > [root at centclust1 mnt]# gluster volume status smbbackup Status
> > > > > of
> > > > > volume: smbbackup
> > > > > Gluster process Port
> > > > > Online
> > > > > Pid
> > > > > --------------------------------------------------------------
> > > > > --
> > > > > --
> > > > > --
> > > > > --
> > > > > ------
> > > > > --
> > > > > Brick centclust1.tplk.loc:/sicherung/bu 49152 Y
> > > > > 31969
> > > > > Brick centclust2.tplk.loc:/sicherung/bu 49152 Y
> > > > > 2124
> > > > > NFS Server on localhost 2049 Y
> > > > > 31983
> > > > > Self-heal Daemon on localhost N/A Y
> > > > > 31987
> > > > > NFS Server on centclust2 2049 Y
> > > > > 2138
> > > > > Self-heal Daemon on centclust2 N/A Y
> > > > > 2142
> > > > >
> > > > > Task Status of Volume smbbackup
> > > > > --------------------------------------------------------------
> > > > > --
> > > > > --
> > > > > --
> > > > > --
> > > > > ------
> > > > > --
> > > > > There are no active volume tasks
> > > > >
> > > > > [root at centclust2 sicherung]# gluster volume status smbbackup
> > > > > Status of
> > > > > volume: smbbackup
> > > > > Gluster process Port
> > > > > Online
> > > > > Pid
> > > > > --------------------------------------------------------------
> > > > > --
> > > > > --
> > > > > --
> > > > > --
> > > > > ------
> > > > > --
> > > > > Brick centclust1.tplk.loc:/sicherung/bu 49152 Y
> > > > > 31969
> > > > > Brick centclust2.tplk.loc:/sicherung/bu 49152 Y
> > > > > 2124
> > > > > NFS Server on localhost 2049 Y
> > > > > 2138
> > > > > Self-heal Daemon on localhost N/A Y
> > > > > 2142
> > > > > NFS Server on 172.17.2.30 2049 Y
> > > > > 31983
> > > > > Self-heal Daemon on 172.17.2.30 N/A Y
> > > > > 31987
> > > > >
> > > > > Task Status of Volume smbbackup
> > > > > --------------------------------------------------------------
> > > > > --
> > > > > --
> > > > > --
> > > > > --
> > > > > ------
> > > > > --
> > > > > There are no active volume tasks
> > > > >
> > > > > I mounted the vol on both servers with:
> > > > >
> > > > > mount -t glusterfs centclust1.tplk.loc:/smbbackup
> > > > > /mnt/sicherung -o acl mount -t glusterfs
> > > > > centclust2.tplk.loc:/smbbackup /mnt/sicherung -o acl
> > > > >
> > > > > But when I write in /mnt/sicherung the files are not
> > > > > replicated to the other node in anyway!??
> > > > >
> > > > > They rest on the local server in /mnt/sicherung and
> > > > > /sicherung/bu On each node separate:#
> > > > > [root at centclust1 sicherung]# pwd /mnt/sicherung
> > > > >
> > > > > [root at centclust1 sicherung]# touch test.txt
> > > > > [root at centclust1 sicherung]# ls test.txt
> > > > > [root at centclust2 sicherung]# pwd /mnt/sicherung
> > > > > [root at centclust2 sicherung]# ls more.txt
> > > > > [root at centclust1 sicherung]# ls -la /sicherung/bu insgesamt 0
> > > > > drwxr-xr-x. 3 root root 38 29. Jul 15:56 .
> > > > > drwxr-xr-x. 3 root root 15 29. Jul 14:31 ..
> > > > > drw-------. 15 root root 142 29. Jul 15:56 .glusterfs
> > > > > -rw-r--r--. 2 root root 0 29. Jul 15:56 test.txt
> > > > > [root at centclust2 sicherung]# ls -la /sicherung/bu insgesamt 0
> > > > > drwxr-xr-x. 3 root root 38 29. Jul 15:32 .
> > > > > drwxr-xr-x. 3 root root 15 29. Jul 14:31 ..
> > > > > drw-------. 7 root root 70 29. Jul 15:32 .glusterfs -rw-r--r--.
> > > > > 2 root root 0 29. Jul 15:32 more.txt
> > > > >
> > > > >
> > > > >
> > > > > Greetings
> > > > > Daniel
> > > > >
> > > > >
> > > > >
> > > > > EDV Daniel Müller
> > > > >
> > > > > Leitung EDV
> > > > > Tropenklinik Paul-Lechler-Krankenhaus Paul-Lechler-Str. 24
> > > > > 72076 Tübingen
> > > > > Tel.: 07071/206-463, Fax: 07071/206-499
> > > > > eMail: mueller at tropenklinik.de
> > > > > Internet: www.tropenklinik.de
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > _______________________________________________
> > > > > Gluster-users mailing list
> > > > > Gluster-users at gluster.org
> > > > > http://supercolony.gluster.org/mailman/listinfo/gluster-users
> > > > >
> > > >
> > > > _______________________________________________
> > > > Gluster-users mailing list
> > > > Gluster-users at gluster.org
> > > > http://supercolony.gluster.org/mailman/listinfo/gluster-users
> > >
> > > _______________________________________________
> > > Gluster-users mailing list
> > > Gluster-users at gluster.org
> > > http://supercolony.gluster.org/mailman/listinfo/gluster-users
> >
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
More information about the Gluster-users
mailing list