[Gluster-users] geo-replication unprivileged user error

Saravanakumar Arumugam sarumuga at redhat.com
Thu Mar 31 06:33:48 UTC 2016


Hi,
Replies inline.

Thanks,
Saravana

On 03/31/2016 04:00 AM, Gmail wrote:
> I’ve rebuilt the cluster again, making a fresh installation. And now 
> the error is different.
>
>
>
>
>
> MASTER NODE MASTER VOL    MASTER BRICK              SLAVE USER    
> SLAVE                               SLAVE NODE          STATUS CRAWL 
> STATUS    LAST_SYNCED
> -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
> master-host01.me.com <http://master-host01.me.com>   geotest       
> /gpool/brick03/geotest    guser guser at slave-host01::geotestdr    N/A 
> Faulty     N/A             N/A
> master-host02.me.com <http://master-host02.me.com>   geotest       
> /gpool/brick03/geotest    guser guser at slave-host01::geotestdr    
> slave-host01 Passive   N/A             N/A
> master-host03.me.com <http://master-host03.me.com>   geotest       
> /gpool/brick03/geotest    guser guser at slave-host01::geotestdr    
> slave-host03 Passive   N/A             N/A
>
There seems to issue with geo-rep setup.

  - All the master bricks seems same..which should not be the case.

What type of volume is this?
Can you get "gluster volume status" and "gluster volume info"  for both 
master and slave volume?

Also, share all the commands you execute to setup this georep session ?

>
>
>
>
>
> [2016-03-30 22:09:31.326898] I [monitor(monitor):221:monitor] Monitor: 
> ------------------------------------------------------------
> [2016-03-30 22:09:31.327461] I [monitor(monitor):222:monitor] Monitor: 
> starting gsyncd worker
> [2016-03-30 22:09:31.544631] I 
> [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: 
> gluster://localhost:geotest -> 
> ssh://guser@slave-host02:gluster://localhost:geotestdr
> [2016-03-30 22:09:31.547542] I [changelogagent(agent):75:__init__] 
> ChangelogAgent: Agent listining...
> [2016-03-30 22:09:31.830554] E 
> [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: 
> connection to peer is broken
> [2016-03-30 22:09:31.831017] W 
> [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: 
> !!!!!!!!!!!!!
> [2016-03-30 22:09:31.831258] W 
> [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: 
> !!! getting "No such file or directory" errors is most likely due to 
> MISCONFIGURATION, please consult 
> https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html
> [2016-03-30 22:09:31.831502] W 
> [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: 
> !!!!!!!!!!!!!
> [2016-03-30 22:09:31.836395] E 
> [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh 
> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i 
> /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S 
> /tmp/gsyncd-aux-ssh-SfXvbB/de372ce5774b5d259c58c5c9522ffc8f.sock 
> guser at slave-host02 /nonexistent/gsyncd --session-owner 
> ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 
> gluster://localhost:geotestdr" returned with 127, saying:
> [2016-03-30 22:09:31.836694] E 
> [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: 
> /nonexistent/gsyncd: No such file or directory
> [2016-03-30 22:09:31.837193] I 
> [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting.
> [2016-03-30 22:09:31.840569] I [repce(agent):92:service_loop] 
> RepceServer: terminating on reaching EOF.
> [2016-03-30 22:09:31.840993] I [syncdutils(agent):220:finalize] <top>: 
> exiting.
> [2016-03-30 22:09:31.840742] I [monitor(monitor):274:monitor] Monitor: 
> worker(/gpool/brick03/geotest) died before establishing connection
> [2016-03-30 22:09:42.130866] I [monitor(monitor):221:monitor] Monitor: 
> ------------------------------------------------------------
> [2016-03-30 22:09:42.131448] I [monitor(monitor):222:monitor] Monitor: 
> starting gsyncd worker
> [2016-03-30 22:09:42.348165] I 
> [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: 
> gluster://localhost:geotest -> 
> ssh://guser@slave-host02:gluster://localhost:geotestdr
> [2016-03-30 22:09:42.349118] I [changelogagent(agent):75:__init__] 
> ChangelogAgent: Agent listining...
> [2016-03-30 22:09:42.653141] E 
> [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: 
> connection to peer is broken
> [2016-03-30 22:09:42.653656] W 
> [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: 
> !!!!!!!!!!!!!
> [2016-03-30 22:09:42.653898] W 
> [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: 
> !!! getting "No such file or directory" errors is most likely due to 
> MISCONFIGURATION, please consult 
> https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html
> [2016-03-30 22:09:42.654129] W 
> [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: 
> !!!!!!!!!!!!!
> [2016-03-30 22:09:42.659329] E 
> [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh 
> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i 
> /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S 
> /tmp/gsyncd-aux-ssh-6r8rxx/de372ce5774b5d259c58c5c9522ffc8f.sock 
> guser at slave-host02 /nonexistent/gsyncd --session-owner 
> ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 
> gluster://localhost:geotestdr" returned with 127, saying:
> [2016-03-30 22:09:42.659626] E 
> [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: 
> /nonexistent/gsyncd: No such file or directory
> [2016-03-30 22:09:42.660140] I 
> [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting.
> [2016-03-30 22:09:42.662802] I [repce(agent):92:service_loop] 
> RepceServer: terminating on reaching EOF.
> [2016-03-30 22:09:42.663197] I [syncdutils(agent):220:finalize] <top>: 
> exiting.
> [2016-03-30 22:09:42.663024] I [monitor(monitor):274:monitor] Monitor: 
> worker(/gpool/brick03/geotest) died before establishing connection
>
>
> —Bishoy
>
>> On Mar 30, 2016, at 10:50 AM, Gmail <b.s.mikhael at gmail.com 
>> <mailto:b.s.mikhael at gmail.com>> wrote:
>>
>> I’ve tried changing the permissions to 777 on /var/log/glusterfs on 
>> all the slave nodes, but still no luck :(
>>
>> here is the log from the master node where I created and started the 
>> geo-replication session.
>>
>> [2016-03-30 17:14:53.463150] I [monitor(monitor):221:monitor] 
>> Monitor: ------------------------------------------------------------
>> [2016-03-30 17:14:53.463669] I [monitor(monitor):222:monitor] 
>> Monitor: starting gsyncd worker
>> [2016-03-30 17:14:53.603774] I [changelogagent(agent):75:__init__] 
>> ChangelogAgent: Agent listining...
>> [2016-03-30 17:14:53.604080] I 
>> [gsyncd(/mnt/brick10/xfsvol2):649:main_i] <top>: syncing: 
>> gluster://localhost:xfsvol2 -> 
>> ssh://guser@slave-host01:gluster://localhost:xfsvol2dr
>> [2016-03-30 17:14:54.210602] E 
>> [syncdutils(/mnt/brick10/xfsvol2):252:log_raise_exception] <top>: 
>> connection to peer is broken
>> [2016-03-30 17:14:54.211117] E 
>> [resource(/mnt/brick10/xfsvol2):222:errlog] Popen: command "ssh 
>> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i 
>> /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S 
>> /tmp/gsyncd-aux-ssh-evONxc/3bda60dc6e900c0833fed4e4fdfbd480.sock 
>> guser at slave-host01 /nonexistent/gsyncd --session-owner 
>> ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 
>> gluster://localhost:xfsvol2dr" returned with 1, saying:
>> [2016-03-30 17:14:54.211376] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 
>> 17:14:53.933174] I [cli.c:720:main] 0-cli: Started running 
>> /usr/sbin/gluster with version 3.7.3
>> [2016-03-30 17:14:54.211631] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 
>> 17:14:53.933225] I [cli.c:608:cli_rpc_init] 0-cli: Connecting to 
>> remote glusterd at localhost
>> [2016-03-30 17:14:54.211828] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 
>> 17:14:54.074207] I [MSGID: 101190] 
>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started 
>> thread with index 1
>> [2016-03-30 17:14:54.212017] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 
>> 17:14:54.074302] I [socket.c:2409:socket_event_handler] 0-transport: 
>> disconnecting now
>> [2016-03-30 17:14:54.212199] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 
>> 17:14:54.077207] I [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: 
>> Received resp to getwd
>> [2016-03-30 17:14:54.212380] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 
>> 17:14:54.077269] I [input.c:36:cli_batch] 0-: Exiting with: 0
>> [2016-03-30 17:14:54.212584] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> ERROR:root:FAIL:
>> [2016-03-30 17:14:54.212774] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> Traceback 
>> (most recent call last):
>> [2016-03-30 17:14:54.212954] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 165, in main
>> [2016-03-30 17:14:54.213131] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>     main_i()
>> [2016-03-30 17:14:54.213308] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 638, in main_i
>> [2016-03-30 17:14:54.213500] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>     
>> startup(go_daemon=go_daemon, log_file=log_file, label=label)
>> [2016-03-30 17:14:54.213690] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 144, in 
>> startup
>> [2016-03-30 17:14:54.213890] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>     
>> GLogger._gsyncd_loginit(**kw)
>> [2016-03-30 17:14:54.214068] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 107, in 
>> _gsyncd_loginit
>> [2016-03-30 17:14:54.214246] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> 
>> cls.setup(label=kw.get('label'), **lkw)
>> [2016-03-30 17:14:54.214422] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, in setup
>> [2016-03-30 17:14:54.214622] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>     
>> logging_handler = handlers.WatchedFileHandler(lprm['filename'])
>> [2016-03-30 17:14:54.214802] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__
>> [2016-03-30 17:14:54.214977] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> 
>> logging.FileHandler.__init__(self, filename, mode, encoding, delay)
>> [2016-03-30 17:14:54.215152] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__
>> [2016-03-30 17:14:54.215327] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>     
>> StreamHandler.__init__(self, self._open())
>> [2016-03-30 17:14:54.215523] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>   File 
>> "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open
>> [2016-03-30 17:14:54.215703] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh>     stream = 
>> open(self.baseFilename, self.mode)
>> [2016-03-30 17:14:54.215883] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> IOError: 
>> [Errno 13] Permission denied: 
>> '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Axfsvol2dr.log'
>> [2016-03-30 17:14:54.216063] E 
>> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> failed with 
>> IOError.
>> [2016-03-30 17:14:54.216500] I 
>> [syncdutils(/mnt/brick10/xfsvol2):220:finalize] <top>: exiting.
>> [2016-03-30 17:14:54.218672] I [repce(agent):92:service_loop] 
>> RepceServer: terminating on reaching EOF.
>> [2016-03-30 17:14:54.219063] I [syncdutils(agent):220:finalize] 
>> <top>: exiting.
>> [2016-03-30 17:14:54.218930] I [monitor(monitor):274:monitor] 
>> Monitor: worker(/mnt/brick10/xfsvol2) died before establishing connection
>>
>> *—Bishoy*
>> *
>> *
>>> On Mar 29, 2016, at 1:05 AM, Aravinda <avishwan at redhat.com 
>>> <mailto:avishwan at redhat.com>> wrote:
>>>
>>> Geo-replication command should be run as privileged user itself.
>>>
>>> gluster volume geo-replication <MASTERVOL> <SLAVEUSER>@<SLAVEHOST> start
>>>
>>> and then check the status, if it shows Faulty then please share the 
>>> log files present in 
>>> /var/log/glusterfs/geo-replication/<MASTERVOL>/*.log
>>>
>>> regards
>>> Aravinda
>>> On 03/29/2016 12:51 PM, Gmail wrote:
>>>> I’ve been trying to setup geo-replication using Gluster 3.7.3 on 
>>>> OEL 6.5
>>>> It keeps giving me faulty session.
>>>> I’ve tried to use root user instead, it works fine!
>>>>
>>>> I’ve followed literally the documentation but no luck getting the 
>>>> unprivileged user working.
>>>>
>>>> I’ve tried running /usr/libexec/glusterfs/gsyncd on the slave node 
>>>> using the unprivileged user, and that’s what I get.
>>>>
>>>> /usr/libexec/glusterfs/gsyncd --session-owner 
>>>> ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 
>>>> gluster://localhost:vol01dr
>>>> [2016-03-29 00:52:49.058244] I [cli.c:720:main] 0-cli: Started 
>>>> running /usr/sbin/gluster with version 3.7.3
>>>> [2016-03-29 00:52:49.058297] I [cli.c:608:cli_rpc_init] 0-cli: 
>>>> Connecting to remote glusterd at localhost
>>>> [2016-03-29 00:52:49.174686] I [MSGID: 101190] 
>>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started 
>>>> thread with index 1
>>>> [2016-03-29 00:52:49.174768] I [socket.c:2409:socket_event_handler] 
>>>> 0-transport: disconnecting now
>>>> [2016-03-29 00:52:49.177482] I 
>>>> [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: Received resp to getwd
>>>> [2016-03-29 00:52:49.177545] I [input.c:36:cli_batch] 0-: Exiting 
>>>> with: 0
>>>> ERROR:root:FAIL:
>>>> Traceback (most recent call last):
>>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 
>>>> 165, in main
>>>>   main_i()
>>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 
>>>> 638, in main_i
>>>>   startup(go_daemon=go_daemon, log_file=log_file, label=label)
>>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 
>>>> 144, in startup
>>>>   GLogger._gsyncd_loginit(**kw)
>>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 
>>>> 107, in _gsyncd_loginit
>>>>   cls.setup(label=kw.get('label'), **lkw)
>>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, 
>>>> in setup
>>>>   logging_handler = handlers.WatchedFileHandler(lprm['filename'])
>>>> File "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__
>>>>   logging.FileHandler.__init__(self, filename, mode, encoding, delay)
>>>> File "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__
>>>>   StreamHandler.__init__(self, self._open())
>>>> File "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open
>>>>   stream = open(self.baseFilename, self.mode)
>>>> IOError: [Errno 13] Permission denied: 
>>>> '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Avol01dr.log'
>>>> failed with IOError.
>>>>
>>>>
>>>> *— Bishoy*
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>
>>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160331/2e3b9df0/attachment.html>


More information about the Gluster-users mailing list