[Gluster-users] What a brick is missing in `sudo gluster volume status`?

Peng Yu pengyu.ut at gmail.com
Sat Mar 22 19:46:47 UTC 2014


Hi,

I solved the problem by deleting the volume and create it again. It
seems that the problem is due to something else and resetting gluster
might be the way to solve the problem. Does anybody know a better way
to completely reset gluster? (Deleting and recreating the volume seems
to be not a good way in a production system.) Thanks.

On Sat, Mar 22, 2014 at 2:14 PM, Peng Yu <pengyu.ut at gmail.com> wrote:
> Hi,
>
> Here are the respective IP addresses of both servers. Why should I
> remove "auth.allow: 172.17.*.*"? (And how to remove it?)
>
> pengy at rigel:~$ ifconfig |grep -A 7 '^br1'
> br1       Link encap:Ethernet  HWaddr c8:1f:66:e2:90:45
>           inet addr:172.17.1.1  Bcast:172.17.255.255  Mask:255.255.0.0
>           inet6 addr: fe80::ca1f:66ff:fee2:9045/64 Scope:Link
>           UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>           RX packets:312191 errors:0 dropped:0 overruns:0 frame:0
>           TX packets:210807 errors:0 dropped:0 overruns:0 carrier:0
>           collisions:0 txqueuelen:0
>           RX bytes:3741197826 (3.7 GB)  TX bytes:25954291 (25.9 MB)
> pengy at betelgeuse:~$  ifconfig |grep -A 7 '^br1'
> br1       Link encap:Ethernet  HWaddr c8:1f:66:df:01:0b
>           inet addr:172.17.2.1  Bcast:172.17.255.255  Mask:255.255.0.0
>           inet6 addr: fe80::ca1f:66ff:fedf:10b/64 Scope:Link
>           UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>           RX packets:197382 errors:0 dropped:0 overruns:0 frame:0
>           TX packets:90443 errors:0 dropped:0 overruns:0 carrier:0
>           collisions:0 txqueuelen:0
>           RX bytes:11914450 (11.9 MB)  TX bytes:10016451 (10.0 MB)
>
>
> Here are are the firewall information. I don't see anything is wrong.
> Do you see anything wrong? Thanks.
>
> pengy at rigel:~$ sudo ufw app list
> Available applications:
>   OpenSSH
> pengy at rigel:~$ sudo ufw status
> Status: inactive
> pengy at rigel:~$ sudo iptables -L
> Chain INPUT (policy ACCEPT)
> target     prot opt source               destination
> ACCEPT     udp  --  anywhere             anywhere             udp dpt:domain
> ACCEPT     tcp  --  anywhere             anywhere             tcp dpt:domain
> ACCEPT     udp  --  anywhere             anywhere             udp dpt:bootps
> ACCEPT     tcp  --  anywhere             anywhere             tcp dpt:bootps
>
> Chain FORWARD (policy ACCEPT)
> target     prot opt source               destination
> ACCEPT     tcp  --  anywhere             192.168.122.200      state
> NEW,RELATED,ESTABLISHED tcp dpt:ssh
> ACCEPT     all  --  anywhere             192.168.122.0/24     ctstate
> RELATED,ESTABLISHED
> ACCEPT     all  --  192.168.122.0/24     anywhere
> ACCEPT     all  --  anywhere             anywhere
> REJECT     all  --  anywhere             anywhere
> reject-with icmp-port-unreachable
> REJECT     all  --  anywhere             anywhere
> reject-with icmp-port-unreachable
>
> Chain OUTPUT (policy ACCEPT)
> target     prot opt source               destination
>
> pengy at betelgeuse:~$ sudo ufw app list
> Available applications:
>   OpenSSH
> pengy at betelgeuse:~$ sudo ufw status
> Status: inactive
> pengy at betelgeuse:~$ sudo iptables -L
> Chain INPUT (policy ACCEPT)
> target     prot opt source               destination
> ACCEPT     udp  --  anywhere             anywhere             udp dpt:domain
> ACCEPT     tcp  --  anywhere             anywhere             tcp dpt:domain
> ACCEPT     udp  --  anywhere             anywhere             udp dpt:bootps
> ACCEPT     tcp  --  anywhere             anywhere             tcp dpt:bootps
>
> Chain FORWARD (policy ACCEPT)
> target     prot opt source               destination
> ACCEPT     all  --  anywhere             192.168.122.0/24     ctstate
> RELATED,ESTABLISHED
> ACCEPT     all  --  192.168.122.0/24     anywhere
> ACCEPT     all  --  anywhere             anywhere
> REJECT     all  --  anywhere             anywhere
> reject-with icmp-port-unreachable
> REJECT     all  --  anywhere             anywhere
> reject-with icmp-port-unreachable
>
> Chain OUTPUT (policy ACCEPT)
> target     prot opt source               destination
>
>
> On Sat, Mar 22, 2014 at 2:01 PM, Carlos Capriotti
> <capriotti.carlos at gmail.com> wrote:
>> One thing that caught my eyes:
>>
>> auth.allow: 172.17.*.*
>>
>> Can you remove that, restart glusterd/the nodes and try again ?
>>
>> Also, do you have firewall/iptables rules enabled ? If yes, consider testing
>> with iptables/firewall disabled.
>>
>>
>>
>>
>> On Sat, Mar 22, 2014 at 7:09 PM, Peng Yu <pengyu.ut at gmail.com> wrote:
>>>
>>> Hi,
>>>
>>> There should be two bricks in the volume "gv". But `sudo gluster
>>> volume status` does not show `betelgeuse:/mnt/raid6/glusterfs_export`.
>>> Does anybody know what is wrong with this? Thanks.
>>>
>>> pengy at rigel:~$ sudo gluster volume status
>>> Status of volume: gv
>>> Gluster process                        Port    Online    Pid
>>>
>>> ------------------------------------------------------------------------------
>>> Brick rigel:/mnt/raid6/glusterfs_export            49152    Y    38971
>>> NFS Server on localhost                    N/A    N    N/A
>>> Self-heal Daemon on localhost                N/A    N    N/A
>>>
>>> There are no active volume tasks
>>> pengy at rigel:~$ sudo gluster volume info
>>>
>>> Volume Name: gv
>>> Type: Replicate
>>> Volume ID: 64754d6c-3736-41d8-afb5-d8071a6a6a07
>>> Status: Started
>>> Number of Bricks: 1 x 2 = 2
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: rigel:/mnt/raid6/glusterfs_export
>>> Brick2: betelgeuse:/mnt/raid6/glusterfs_export
>>> Options Reconfigured:
>>> auth.allow: 172.17.*.*
>>>
>>> --
>>> Regards,
>>> Peng
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
>>
>
>
>
> --
> Regards,
> Peng



-- 
Regards,
Peng



More information about the Gluster-users mailing list