[Gluster-users] volume process does not start - glusterfs is happy with it?

Felix Kölzow felix.koelzow at gmx.de
Wed Jul 1 17:57:22 UTC 2020


Hey,


what about the device mapper? Everything was mount properly during reboot?

It happens to me if the lvm device mapper got a timeout during the reboot

process while mounting the brick itself.


Regards,

Felix

On 01/07/2020 16:46, lejeczek wrote:
>
> On 30/06/2020 11:31, Barak Sason Rofman wrote:
>> Greetings,
>>
>> I'm not sure if that's directly related to your problem,
>> but on a general level, AFAIK, replica-2 vols are not
>> recommended due to split brain possibility:
>> https://docs.gluster.org/en/latest/Administrator%20Guide/Split%20brain%20and%20ways%20to%20deal%20with%20it/
>>
>> It's recommended to either use replica-3 or arbiter Arbiter.
>>
>> Regards,
>>
>> On Tue, Jun 30, 2020 at 1:14 PM lejeczek
>> <peljasz at yahoo.co.uk <mailto:peljasz at yahoo.co.uk>> wrote:
>>
>>      Hi everybody.
>>
>>      I have two peers in the cluster and a 2-replica volume
>>      which seems okey if it was not for one weird bit -
>>      when a peer reboots then on that peer after a reboot I
>>      see:
>>
>>      $ gluster volume status USERs
>>      Status of volume: USERs
>>      Gluster process                             TCP Port
>>      RDMA Port  Online  Pid
>>      ------------------------------------------------------------------------------
>>      Brick swir.direct:/00.STORAGE/2/0-GLUSTER-U
>>      SERs                                        N/A
>>      N/A        N       N/A
>>      Brick dzien.direct:/00.STORAGE/2/0-GLUSTER-
>>      USERs                                       49152
>>      0          Y       57338
>>      Self-heal Daemon on localhost               N/A
>>      N/A        Y       4302
>>      Self-heal Daemon on dzien.direct            N/A
>>      N/A        Y       57359
>>
>>      Task Status of Volume USERs
>>      ------------------------------------------------------------------------------
>>      There are no active volume tasks
>>
>>      I do not suppose it's expected.
>>      On such rebooted node I see:
>>      $ systemctl status -l glusterd
>>      ● glusterd.service - GlusterFS, a clustered
>>      file-system server
>>         Loaded: loaded
>>      (/usr/lib/systemd/system/glusterd.service; enabled;
>>      vendor preset: enabled)
>>        Drop-In: /etc/systemd/system/glusterd.service.d
>>                 └─override.conf
>>         Active: active (running) since Mon 2020-06-29
>>      21:37:36 BST; 13h ago
>>           Docs: man:glusterd(8)
>>        Process: 4071 ExecStart=/usr/sbin/glusterd -p
>>      /var/run/glusterd.pid --log-level $LOG_LEVEL
>>      $GLUSTERD_OPTIONS (code=exited, status>
>>       Main PID: 4086 (glusterd)
>>          Tasks: 20 (limit: 101792)
>>         Memory: 28.9M
>>         CGroup: /system.slice/glusterd.service
>>                 ├─4086 /usr/sbin/glusterd -p
>>      /var/run/glusterd.pid --log-level INFO
>>                 └─4302 /usr/sbin/glusterfs -s localhost
>>      --volfile-id shd/USERs -p
>>      /var/run/gluster/shd/USERs/USERs-shd.pid -l /var/log/g>
>>
>>      Jun 29 21:37:36 swir.private.pawel systemd[1]:
>>      Starting GlusterFS, a clustered file-system server...
>>      Jun 29 21:37:36 swir.private.pawel systemd[1]: Started
>>      GlusterFS, a clustered file-system server.
>>
>>      And I do not see any other apparent problems nor errors.
>>      On that node I manually:
>>      $ systemctl restart glusterd.service
>>      and...
>>
>>      $ gluster volume status USERs
>>      Status of volume: USERs
>>      Gluster process                             TCP Port
>>      RDMA Port  Online  Pid
>>      ------------------------------------------------------------------------------
>>      Brick swir.direct:/00.STORAGE/2/0-GLUSTER-U
>>      SERs                                        49152
>>      0          Y       103225
>>      Brick dzien.direct:/00.STORAGE/2/0-GLUSTER-
>>      USERs                                       49152
>>      0          Y       57338
>>      Self-heal Daemon on localhost               N/A
>>      N/A        Y       103270
>>      Self-heal Daemon on dzien.direct            N/A
>>      N/A        Y       57359
>>
>>      Is not a puzzle??? I'm on glusterfs-7.6-1.el8.x86_64
>>      I hope somebody can share some thoughts.
>>      many thanks, L.
>>
> That cannot be it!? If the root cause of this problem is
> 2-replica volume then it would be a massive cock-up! Then
> 2-volume replica should be banned and forbidden.
>
> I hope some can suggest a way to troubleshoot it.
>
> ps. we all, I presume all, know problems of 2-replica volumes.
>
> many thanks, L.
>
>
>>      ________
>>
>>
>>
>>      Community Meeting Calendar:
>>
>>      Schedule -
>>      Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>      Bridge: https://bluejeans.com/441850968
>>
>>      Gluster-users mailing list
>>      Gluster-users at gluster.org
>>      <mailto:Gluster-users at gluster.org>
>>      https://lists.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>> --
>> *Barak Sason Rofman*
>>
>> Gluster Storage Development
>>
>> Red Hat Israel <https://www.redhat.com/>
>>
>> 34 Jerusalem rd. Ra'anana, 43501
>>
>> bsasonro at redhat.com <mailto:adi at redhat.com>
>>    T: _+972-9-7692304_
>> M: _+972-52-4326355_
>>
>> @RedHat <https://twitter.com/redhat>   Red Hat
>> <https://www.linkedin.com/company/red-hat>  Red Hat
>> <https://www.facebook.com/redhat.il/>
>> <https://red.ht/sig>
>>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users


More information about the Gluster-users mailing list