From bugzilla at redhat.com Sat Jun 1 13:15:36 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 01 Jun 2019 13:15:36 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22797
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 1 13:15:38 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 01 Jun 2019 13:15:38 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #674 from Worker Ant ---
REVIEW: https://review.gluster.org/22797 (glusterd: remove trivial conditions)
posted (#1) for review on master by Sanju Rakonde
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 1 17:27:55 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 01 Jun 2019 17:27:55 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #675 from Worker Ant ---
REVIEW: https://review.gluster.org/22797 (glusterd: remove trivial conditions)
merged (#1) on master by Sanju Rakonde
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 1 20:20:30 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 01 Jun 2019 20:20:30 +0000
Subject: [Bugs] [Bug 1703948] Self-heal daemon resources are not cleaned
properly after a ec fini
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703948
Mohammed Rafi KC changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|CLOSED |POST
Resolution|NEXTRELEASE |---
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sat Jun 1 20:23:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 01 Jun 2019 20:23:29 +0000
Subject: [Bugs] [Bug 1703948] Self-heal daemon resources are not cleaned
properly after a ec fini
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703948
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22798
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sat Jun 1 20:23:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 01 Jun 2019 20:23:29 +0000
Subject: [Bugs] [Bug 1703948] Self-heal daemon resources are not cleaned
properly after a ec fini
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703948
--- Comment #5 from Worker Ant ---
REVIEW: https://review.gluster.org/22798 (ec/fini: Fix race between xlator
cleanup and on going async fop) posted (#1) for review on master by mohammed
rafi kc
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sat Jun 1 21:01:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 01 Jun 2019 21:01:10 +0000
Subject: [Bugs] [Bug 1716097] New: infra: create
suse-packing@lists.nfs-ganesha.org alias
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716097
Bug ID: 1716097
Summary: infra: create suse-packing at lists.nfs-ganesha.org alias
Product: GlusterFS
Version: mainline
Status: NEW
Component: project-infrastructure
Assignee: bugs at gluster.org
Reporter: kkeithle at redhat.com
CC: bugs at gluster.org, gluster-infra at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
Is there an OSAS ticketing system to use instead of this?
Anyway, forwarded to me.
Thanks
Version-Release number of selected component (if applicable):
How reproducible:
Steps to Reproduce:
1.
2.
3.
Actual results:
Expected results:
Additional info:
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sun Jun 2 09:22:05 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sun, 02 Jun 2019 09:22:05 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22799
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sun Jun 2 09:22:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sun, 02 Jun 2019 09:22:06 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
--- Comment #50 from Worker Ant ---
REVIEW: https://review.gluster.org/22799 (lcov: run more fops on translators)
posted (#1) for review on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sun Jun 2 18:18:13 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sun, 02 Jun 2019 18:18:13 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22800
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sun Jun 2 18:18:13 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sun, 02 Jun 2019 18:18:13 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #676 from Worker Ant ---
REVIEW: https://review.gluster.org/22800 ([WIP] (multiple files) CALLOC ->
MALLOC when serializing a dictionary) posted (#1) for review on master by Yaniv
Kaul
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 02:59:51 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 02:59:51 +0000
Subject: [Bugs] [Bug 1651445] [RFE] storage.reserve option should take size
of disk as input instead of percentage
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1651445
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-03 02:59:51
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/21686 (posix: add storage.reserve-size
option) merged (#13) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 04:01:18 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 04:01:18 +0000
Subject: [Bugs] [Bug 789278] Issues reported by Coverity static analysis tool
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=789278
--- Comment #1631 from Worker Ant ---
REVIEW: https://review.gluster.org/22741 (across: coverity fixes) merged (#12)
on master by Amar Tumballi
--
You are receiving this mail because:
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 04:08:31 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 04:08:31 +0000
Subject: [Bugs] [Bug 1715012] Failure when glusterd is configured to bind
specific IPv6 address. If bind-address is IPv6,
*addr_len will be non-zero and it goes to ret = -1 branch,
which will cause listen failure eventually
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1715012
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-03 04:08:31
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22787 (If bind-address is IPv6 return it
successfully) merged (#2) on release-6 by Sunny Kumar
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 04:08:53 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 04:08:53 +0000
Subject: [Bugs] [Bug 1714172] ec ignores lock contention notifications for
partially acquired locks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714172
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-03 04:08:53
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22774 (cluster/ec: honor contention
notifications for partially acquired locks) merged (#2) on release-6 by Amar
Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 04:23:03 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 04:23:03 +0000
Subject: [Bugs] [Bug 789278] Issues reported by Coverity static analysis tool
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=789278
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22801
--
You are receiving this mail because:
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 04:23:05 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 04:23:05 +0000
Subject: [Bugs] [Bug 789278] Issues reported by Coverity static analysis tool
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=789278
--- Comment #1632 from Worker Ant ---
REVIEW: https://review.gluster.org/22801 (glusterd: coverity fix) posted (#1)
for review on master by MOHIT AGRAWAL
--
You are receiving this mail because:
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 06:22:17 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 06:22:17 +0000
Subject: [Bugs] [Bug 1703322] Need to document about fips-mode-rchecksum in
gluster-7 release notes.
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703322
Yaniv Kaul changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags| |needinfo?(ravishankar at redha
| |t.com)
Severity|unspecified |medium
--- Comment #1 from Yaniv Kaul ---
https://review.gluster.org/#/c/glusterfs/+/22609/ is merged. Can we now
document it?
When can we remove this option altogether and have it as a default (and then
remove all the gf_rsync_md5_checksum() code and friends) ?
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 07:57:57 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 07:57:57 +0000
Subject: [Bugs] [Bug 1714851] issues with 'list.h' elements in clang-scan
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714851
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |jahernan at redhat.com
--- Comment #1 from Xavi Hernandez ---
I'm not sure we really have an issue in list_for_each_entry_safe(). Even if the
list is empty and list_first_entry() is used (which is true that it returns a
bad pointer when list is empty), what we get is a pointer to an invalid
structure. That's true. However, the macro only dereferences the 'list' field,
which is guaranteed to be valid, even if the list is empty, and in this case it
will exit the loop, so no unsafe pointers will be passed to the body of the
loop.
Additionally, clang-scan complains about the entry pointer being NULL inside
the loop. The only case where this can happen is when the list is not
initialized with INIT_LIST_HEAD() and the memory is cleared with 0's. However
clang-scan doesn't provide a trace path from allocation to
list_for_each_entry_safe() call where this can be proved. So my guess is that
clang-scan assumes that any value is possible for a given pointer passed as an
argument. In that case many false-positives will appear, since it's assuming
something that is not true most of the cases.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 08:26:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 08:26:06 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22803
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 08:26:07 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 08:26:07 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
--- Comment #51 from Worker Ant ---
REVIEW: https://review.gluster.org/22803 (tests/geo-rep: Add geo-rep glusterd
test cases) posted (#1) for review on master by Kotresh HR
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 08:43:47 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 08:43:47 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
--- Comment #52 from Worker Ant ---
REVIEW: https://review.gluster.org/22789 (lcov: improve line coverage) merged
(#2) on master by Xavi Hernandez
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 08:51:32 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 08:51:32 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22804
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 08:51:33 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 08:51:33 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #677 from Worker Ant ---
REVIEW: https://review.gluster.org/22804 (tests/geo-rep: Fix the comment)
posted (#1) for review on master by Kotresh HR
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 13:10:37 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 13:10:37 +0000
Subject: [Bugs] [Bug 1712668] Remove-brick shows warning
cluster.force-migration enabled where as cluster.force-migration is
disabled on the volume
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1712668
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22805
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 13:10:39 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 13:10:39 +0000
Subject: [Bugs] [Bug 1712668] Remove-brick shows warning
cluster.force-migration enabled where as cluster.force-migration is
disabled on the volume
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1712668
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22805 (cli: Remove-brick warning seems
unnecessary) posted (#1) for review on master by Shwetha K Acharya
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 13:31:57 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 13:31:57 +0000
Subject: [Bugs] [Bug 1597798] 'mv' of directory on encrypted volume fails
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1597798
Yaniv Kaul changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags| |needinfo?(vbellur at redhat.co
| |m)
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 13:32:18 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 13:32:18 +0000
Subject: [Bugs] [Bug 1648169] Fuse mount would crash if features.encryption
is on in the version from 3.13.0 to 4.1.5
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1648169
Yaniv Kaul changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags| |needinfo?(vbellur at redhat.co
| |m)
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 13:32:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 13:32:29 +0000
Subject: [Bugs] [Bug 1714973] upgrade after tier code removal results in
peer rejection.
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714973
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-03 13:32:29
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22785 (glusterd/tier: gluster upgrade broken
because of tier) merged (#5) on master by Atin Mukherjee
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 13:33:41 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 13:33:41 +0000
Subject: [Bugs] [Bug 1705351] glusterfsd crash after days of running
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1705351
Yaniv Kaul changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags| |needinfo?(jahernan at redhat.c
| |om)
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 13:34:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 13:34:29 +0000
Subject: [Bugs] [Bug 1635784] brick process segfault
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1635784
Yaniv Kaul changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |CLOSED
Resolution|--- |INSUFFICIENT_DATA
Last Closed| |2019-06-03 13:34:29
--- Comment #7 from Yaniv Kaul ---
(In reply to Yaniv Kaul from comment #6)
> Does it still happen on newer releases?
Closing for the time being. Please re-open if you have more information.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 14:05:13 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 14:05:13 +0000
Subject: [Bugs] [Bug 1703322] Need to document about fips-mode-rchecksum in
gluster-7 release notes.
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703322
Ravishankar N changed:
What |Removed |Added
----------------------------------------------------------------------------
Version|4.1 |mainline
Flags|needinfo?(ravishankar at redha |
|t.com) |
--- Comment #2 from Ravishankar N ---
(In reply to Yaniv Kaul from comment #1)
> https://review.gluster.org/#/c/glusterfs/+/22609/ is merged. Can we now
> document it?
>
I was targeting it for the glusterfs-7 release notes.
> When can we remove this option altogether and have it as a default (and then
> remove all the gf_rsync_md5_checksum() code and friends) ?
Technically, we could do it today since 3.x is EOL and 4.1 onwards clients have
the logic to check the dict for what type of checksum the server is sending and
act accordingly. But people might use 3.x clients with 4.x or later servers 'as
long as the mount succeeds', so maybe it is better to have it for some more
time.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 14:06:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 14:06:06 +0000
Subject: [Bugs] [Bug 1716440] New: SMBD thread panics when connected to from
OS X machine
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716440
Bug ID: 1716440
Summary: SMBD thread panics when connected to from OS X machine
Product: GlusterFS
Version: 6
Hardware: x86_64
OS: Linux
Status: NEW
Component: libgfapi
Severity: high
Assignee: bugs at gluster.org
Reporter: ryan at magenta.tv
QA Contact: bugs at gluster.org
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Created attachment 1576680
--> https://bugzilla.redhat.com/attachment.cgi?id=1576680&action=edit
Debug level 10 log of client connection when panic occurs
Description of problem:
When connecting to a share, the SMB thread for that client panics and
constantly restarts. This was tested from a machine running OS X 10.14.4. I've
not been able to test from a windows machine yet.
Version-Release number of selected component (if applicable):
Gluster = 6.1
Samba = 4.9.6
How reproducible:
Every time
SMB configuration:
[global]
security = user
netbios name = NAS01
clustering = no
server signing = no
max log size = 10000
log file = /var/log/samba/log-%M-test.smbd
logging = file
log level = 10
passdb backend = tdbsam
guest account = nobody
map to guest = bad user
force directory mode = 0777
force create mode = 0777
create mask = 0777
directory mask = 0777
store dos attributes = yes
load printers = no
printing = bsd
printcap name = /dev/null
disable spoolss = yes
glusterfs:volfile_server = localhost
kernel share modes = No
[VFS]
vfs objects = glusterfs
glusterfs:volume = mcv02
path = /
read only = no
guest ok = yes
Steps to Reproduce:
1. Use provided SMB configuration
2. Restart SMB service
3. Connect to share from client using guest user
4. Tail client logs on server to see panics
Actual results:
SMB thread panics and restarts
Expected results:
Client connects and SMB thread doesn't panic
Additional info:
Tested without Gluster VFS and used the FUSE mount point instead and system did
not panic
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 14:08:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 14:08:10 +0000
Subject: [Bugs] [Bug 1663519] Memory leak when smb.conf has "store dos
attributes = yes"
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1663519
ryan at magenta.tv changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags|needinfo?(ryan at magenta.tv) |
--- Comment #5 from ryan at magenta.tv ---
Hi Anoop,
Sorry for the delay.
I've tried to re-test, however we're now using Gluster 6.1 and Samba 4.9.6.
Another issue has come up which is preventing me testing this issue.
I've raised a bug for it here
https://bugzilla.redhat.com/show_bug.cgi?id=1716440.
Once i'm able to re-test I will update this ticket.
Best,
Ryan
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 14:09:07 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 14:09:07 +0000
Subject: [Bugs] [Bug 1716440] SMBD thread panics when connected to from OS X
machine
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716440
ryan at magenta.tv changed:
What |Removed |Added
----------------------------------------------------------------------------
Component|libgfapi |gluster-smb
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 14:13:43 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 14:13:43 +0000
Subject: [Bugs] [Bug 1709248] [geo-rep]: Non-root - Unable to set up
mountbroker root directory and group
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1709248
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |amukherj at redhat.com
Blocks| |1708043
Depends On|1708043 |
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1708043
[Bug 1708043] [geo-rep]: Non-root - Unable to set up mountbroker root directory
and group
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 14:33:05 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 14:33:05 +0000
Subject: [Bugs] [Bug 1716455] New: OS X error -50 when creating sub-folder
on Samba share when using Gluster VFS
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716455
Bug ID: 1716455
Summary: OS X error -50 when creating sub-folder on Samba share
when using Gluster VFS
Product: GlusterFS
Version: 6
Hardware: x86_64
OS: Mac OS
Status: NEW
Component: gluster-smb
Severity: high
Assignee: bugs at gluster.org
Reporter: ryan at magenta.tv
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Created attachment 1576693
--> https://bugzilla.redhat.com/attachment.cgi?id=1576693&action=edit
Debug level 10 log
Description of problem:
OS X finder produces -50 error when trying to create a folder anywhere other
than at the top of a share. This occurs when using the Gluster VFS module.
Version-Release number of selected component (if applicable):
OS X = 10.14.4
Samba = 4.9.6
Gluster = 6.1
How reproducible:
Everytime
Steps to Reproduce:
1. Connect to share
2. Create folder at root of share
3. Go into that folder
4. Try to create folder
5. Create fails and produces error -50
Actual results:
Error -50 produced and folder is not created
Expected results:
Folder is created without error
Additional info:
SMB configuration:
[global]
security = ADS
workgroup = DOMAIN
realm = DOMAIN.LOCAL
netbios name = NAS01
max protocol = SMB3
min protocol = SMB2
ea support = yes
clustering = yes
server signing = no
max log size = 10000
glusterfs:loglevel = 5
log file = /var/log/samba/log-%M.smbd
logging = file
log level = 10
template shell = /sbin/nologin
winbind offline logon = false
winbind refresh tickets = yes
winbind enum users = Yes
winbind enum groups = Yes
allow trusted domains = yes
passdb backend = tdbsam
idmap cache time = 604800
idmap negative cache time = 300
winbind cache time = 604800
idmap config magenta:backend = rid
idmap config magenta:range = 10000-999999
idmap config * : backend = tdb
idmap config * : range = 3000-7999
guest account = nobody
map to guest = bad user
force directory mode = 0777
force create mode = 0777
create mask = 0777
directory mask = 0777
hide unreadable = no
store dos attributes = no
unix extensions = no
load printers = no
printing = bsd
printcap name = /dev/null
disable spoolss = yes
glusterfs:volfile_server = localhost
kernel share modes = No
strict locking = auto
oplocks = yes
durable handles = yes
kernel oplocks = no
posix locking = no
level2 oplocks = no
readdir_attr:aapl_rsize = yes
readdir_attr:aapl_finder_info = no
readdir_attr:aapl_max_access = no
[qc_only]
guest ok = no
read only = no
vfs objects = glusterfs
glusterfs:volume = mcv01
path = "/data/qc_only"
valid users = @"QC_ops"
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 19:22:48 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 19:22:48 +0000
Subject: [Bugs] [Bug 1716626] New: Invalid memory access while executing
cleanup_and_exit
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716626
Bug ID: 1716626
Summary: Invalid memory access while executing cleanup_and_exit
Product: Red Hat Gluster Storage
Version: rhgs-3.5
Status: NEW
Component: replicate
Keywords: Reopened
Assignee: ksubrahm at redhat.com
Reporter: rkavunga at redhat.com
QA Contact: nchilaka at redhat.com
CC: bugs at gluster.org, pkarampu at redhat.com,
rhs-bugs at redhat.com, sankarshan at redhat.com,
storage-qa-internal at redhat.com
Depends On: 1708926
Target Milestone: ---
Classification: Red Hat
+++ This bug was initially created as a clone of Bug #1708926 +++
Description of problem:
when executing a cleanup_and_exit, a shd daemon is crashed. This is because
there is a chance that a parallel graph free thread might be executing another
cleanup
Version-Release number of selected component (if applicable):
How reproducible:
Steps to Reproduce:
1. run ./tests/bugs/glusterd/reset-brick-and-daemons-follow-quorum.t in a loop
2.
3.
Actual results:
Expected results:
Additional info:
--- Additional comment from Worker Ant on 2019-05-11 17:59:31 UTC ---
REVIEW: https://review.gluster.org/22709 (glusterfsd/cleanup: Protect graph
object under a lock) posted (#1) for review on master by mohammed rafi kc
--- Additional comment from Pranith Kumar K on 2019-05-14 07:09:23 UTC ---
Rafi,
Could you share the bt of the core so that it is easier to understand why
exactly it crashed?
Pranith
--- Additional comment from Mohammed Rafi KC on 2019-05-14 16:01:36 UTC ---
Stack trace of thread 30877:
#0 0x0000000000406a07 cleanup_and_exit (glusterfsd)
#1 0x0000000000406b5d glusterfs_sigwaiter (glusterfsd)
#2 0x00007f51000cd58e start_thread (libpthread.so.0)
#3 0x00007f50ffd1d683 __clone (libc.so.6)
Stack trace of thread 30879:
#0 0x00007f51000d3a7a futex_abstimed_wait_cancelable
(libpthread.so.0)
#1 0x00007f51003b8616 syncenv_task (libglusterfs.so.0)
#2 0x00007f51003b9240 syncenv_processor (libglusterfs.so.0)
#3 0x00007f51000cd58e start_thread (libpthread.so.0)
#4 0x00007f50ffd1d683 __clone (libc.so.6)
Stack trace of thread 30881:
#0 0x00007f50ffd14cdf __GI___select (libc.so.6)
#1 0x00007f51003ef1cd runner (libglusterfs.so.0)
#2 0x00007f51000cd58e start_thread (libpthread.so.0)
#3 0x00007f50ffd1d683 __clone (libc.so.6)
Stack trace of thread 30880:
#0 0x00007f51000d3a7a futex_abstimed_wait_cancelable
(libpthread.so.0)
#1 0x00007f51003b8616 syncenv_task (libglusterfs.so.0)
#2 0x00007f51003b9240 syncenv_processor (libglusterfs.so.0)
#3 0x00007f51000cd58e start_thread (libpthread.so.0)
#4 0x00007f50ffd1d683 __clone (libc.so.6)
Stack trace of thread 30876:
#0 0x00007f51000d7500 __GI___nanosleep (libpthread.so.0)
#1 0x00007f510038a346 gf_timer_proc (libglusterfs.so.0)
#2 0x00007f51000cd58e start_thread (libpthread.so.0)
#3 0x00007f50ffd1d683 __clone (libc.so.6)
Stack trace of thread 30882:
#0 0x00007f50ffd1e06e epoll_ctl (libc.so.6)
#1 0x00007f51003d931e event_handled_epoll (libglusterfs.so.0)
#2 0x00007f50eed9a781 socket_event_poll_in (socket.so)
#3 0x00007f51003d8c9b event_dispatch_epoll_handler
(libglusterfs.so.0)
#4 0x00007f51000cd58e start_thread (libpthread.so.0)
#5 0x00007f50ffd1d683 __clone (libc.so.6)
Stack trace of thread 30875:
#0 0x00007f51000cea6d __GI___pthread_timedjoin_ex
(libpthread.so.0)
#1 0x00007f51003d8387 event_dispatch_epoll (libglusterfs.so.0)
#2 0x0000000000406592 main (glusterfsd)
#3 0x00007f50ffc44413 __libc_start_main (libc.so.6)
#4 0x00000000004067de _start (glusterfsd)
Stack trace of thread 30878:
#0 0x00007f50ffce97f8 __GI___nanosleep (libc.so.6)
#1 0x00007f50ffce96fe __sleep (libc.so.6)
#2 0x00007f51003a4f5a pool_sweeper (libglusterfs.so.0)
#3 0x00007f51000cd58e start_thread (libpthread.so.0)
#4 0x00007f50ffd1d683 __clone (libc.so.6)
Stack trace of thread 30883:
#0 0x00007f51000d6b8d __lll_lock_wait (libpthread.so.0)
#1 0x00007f51000cfda9 __GI___pthread_mutex_lock
(libpthread.so.0)
#2 0x00007f510037cd1f _gf_msg_plain_internal
(libglusterfs.so.0)
#3 0x00007f510037ceb3 _gf_msg_plain (libglusterfs.so.0)
#4 0x00007f5100382d43 gf_log_dump_graph (libglusterfs.so.0)
#5 0x00007f51003b514f glusterfs_process_svc_attach_volfp
(libglusterfs.so.0)
#6 0x000000000040b16d mgmt_process_volfile (glusterfsd)
#7 0x0000000000410792 mgmt_getspec_cbk (glusterfsd)
#8 0x00007f51003256b1 rpc_clnt_handle_reply (libgfrpc.so.0)
#9 0x00007f5100325a53 rpc_clnt_notify (libgfrpc.so.0)
#10 0x00007f5100322973 rpc_transport_notify (libgfrpc.so.0)
#11 0x00007f50eed9a45c socket_event_poll_in (socket.so)
#12 0x00007f51003d8c9b event_dispatch_epoll_handler
(libglusterfs.so.0)
#13 0x00007f51000cd58e start_thread (libpthread.so.0)
#14 0x00007f50ffd1d683 __clone (libc.so.6)
--- Additional comment from Pranith Kumar K on 2019-05-15 05:34:33 UTC ---
(In reply to Mohammed Rafi KC from comment #3)
> Stack trace of thread 30877:
> #0 0x0000000000406a07 cleanup_and_exit (glusterfsd)
> #1 0x0000000000406b5d glusterfs_sigwaiter (glusterfsd)
> #2 0x00007f51000cd58e start_thread (libpthread.so.0)
> #3 0x00007f50ffd1d683 __clone (libc.so.6)
>
> Stack trace of thread 30879:
> #0 0x00007f51000d3a7a futex_abstimed_wait_cancelable
> (libpthread.so.0)
> #1 0x00007f51003b8616 syncenv_task (libglusterfs.so.0)
> #2 0x00007f51003b9240 syncenv_processor (libglusterfs.so.0)
> #3 0x00007f51000cd58e start_thread (libpthread.so.0)
> #4 0x00007f50ffd1d683 __clone (libc.so.6)
>
> Stack trace of thread 30881:
> #0 0x00007f50ffd14cdf __GI___select (libc.so.6)
> #1 0x00007f51003ef1cd runner (libglusterfs.so.0)
> #2 0x00007f51000cd58e start_thread (libpthread.so.0)
> #3 0x00007f50ffd1d683 __clone (libc.so.6)
>
> Stack trace of thread 30880:
> #0 0x00007f51000d3a7a futex_abstimed_wait_cancelable
> (libpthread.so.0)
> #1 0x00007f51003b8616 syncenv_task (libglusterfs.so.0)
> #2 0x00007f51003b9240 syncenv_processor (libglusterfs.so.0)
> #3 0x00007f51000cd58e start_thread (libpthread.so.0)
> #4 0x00007f50ffd1d683 __clone (libc.so.6)
>
> Stack trace of thread 30876:
> #0 0x00007f51000d7500 __GI___nanosleep (libpthread.so.0)
> #1 0x00007f510038a346 gf_timer_proc (libglusterfs.so.0)
> #2 0x00007f51000cd58e start_thread (libpthread.so.0)
> #3 0x00007f50ffd1d683 __clone (libc.so.6)
>
> Stack trace of thread 30882:
> #0 0x00007f50ffd1e06e epoll_ctl (libc.so.6)
> #1 0x00007f51003d931e event_handled_epoll
> (libglusterfs.so.0)
> #2 0x00007f50eed9a781 socket_event_poll_in (socket.so)
> #3 0x00007f51003d8c9b event_dispatch_epoll_handler
> (libglusterfs.so.0)
> #4 0x00007f51000cd58e start_thread (libpthread.so.0)
> #5 0x00007f50ffd1d683 __clone (libc.so.6)
>
> Stack trace of thread 30875:
> #0 0x00007f51000cea6d __GI___pthread_timedjoin_ex
> (libpthread.so.0)
> #1 0x00007f51003d8387 event_dispatch_epoll
> (libglusterfs.so.0)
> #2 0x0000000000406592 main (glusterfsd)
> #3 0x00007f50ffc44413 __libc_start_main (libc.so.6)
> #4 0x00000000004067de _start (glusterfsd)
>
> Stack trace of thread 30878:
> #0 0x00007f50ffce97f8 __GI___nanosleep (libc.so.6)
> #1 0x00007f50ffce96fe __sleep (libc.so.6)
> #2 0x00007f51003a4f5a pool_sweeper (libglusterfs.so.0)
> #3 0x00007f51000cd58e start_thread (libpthread.so.0)
> #4 0x00007f50ffd1d683 __clone (libc.so.6)
>
> Stack trace of thread 30883:
> #0 0x00007f51000d6b8d __lll_lock_wait (libpthread.so.0)
> #1 0x00007f51000cfda9 __GI___pthread_mutex_lock
> (libpthread.so.0)
> #2 0x00007f510037cd1f _gf_msg_plain_internal
> (libglusterfs.so.0)
> #3 0x00007f510037ceb3 _gf_msg_plain (libglusterfs.so.0)
> #4 0x00007f5100382d43 gf_log_dump_graph (libglusterfs.so.0)
> #5 0x00007f51003b514f glusterfs_process_svc_attach_volfp
> (libglusterfs.so.0)
> #6 0x000000000040b16d mgmt_process_volfile (glusterfsd)
> #7 0x0000000000410792 mgmt_getspec_cbk (glusterfsd)
> #8 0x00007f51003256b1 rpc_clnt_handle_reply (libgfrpc.so.0)
> #9 0x00007f5100325a53 rpc_clnt_notify (libgfrpc.so.0)
> #10 0x00007f5100322973 rpc_transport_notify (libgfrpc.so.0)
> #11 0x00007f50eed9a45c socket_event_poll_in (socket.so)
> #12 0x00007f51003d8c9b event_dispatch_epoll_handler
> (libglusterfs.so.0)
> #13 0x00007f51000cd58e start_thread (libpthread.so.0)
> #14 0x00007f50ffd1d683 __clone (libc.so.6)
Was graph->active NULL? What lead to the crash?
--- Additional comment from Worker Ant on 2019-05-17 18:08:44 UTC ---
REVIEW: https://review.gluster.org/22743 (afr/frame: Destroy frame after
afr_selfheal_entry_granular) posted (#1) for review on master by mohammed rafi
kc
--- Additional comment from Worker Ant on 2019-05-21 11:37:12 UTC ---
REVIEW: https://review.gluster.org/22743 (afr/frame: Destroy frame after
afr_selfheal_entry_granular) merged (#3) on master by Pranith Kumar Karampuri
--- Additional comment from Worker Ant on 2019-05-31 11:28:15 UTC ---
REVIEW: https://review.gluster.org/22709 (glusterfsd/cleanup: Protect graph
object under a lock) merged (#10) on master by Amar Tumballi
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1708926
[Bug 1708926] Invalid memory access while executing cleanup_and_exit
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 19:22:48 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 19:22:48 +0000
Subject: [Bugs] [Bug 1708926] Invalid memory access while executing
cleanup_and_exit
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1708926
Mohammed Rafi KC changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1716626
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1716626
[Bug 1716626] Invalid memory access while executing cleanup_and_exit
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Mon Jun 3 19:22:50 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 19:22:50 +0000
Subject: [Bugs] [Bug 1716626] Invalid memory access while executing
cleanup_and_exit
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716626
RHEL Product and Program Management changed:
What |Removed |Added
----------------------------------------------------------------------------
Rule Engine Rule| |Gluster: set proposed
| |release flag for new BZs at
| |RHGS
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Mon Jun 3 19:23:36 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Mon, 03 Jun 2019 19:23:36 +0000
Subject: [Bugs] [Bug 1716626] Invalid memory access while executing
cleanup_and_exit
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716626
Mohammed Rafi KC changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
Assignee|ksubrahm at redhat.com |rkavunga at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 00:07:30 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 00:07:30 +0000
Subject: [Bugs] [Bug 1716695] New: Fix memory leaks that are present even
after an xlator fini [client side xlator]
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716695
Bug ID: 1716695
Summary: Fix memory leaks that are present even after an xlator
fini [client side xlator]
Product: GlusterFS
Version: mainline
Status: NEW
Component: core
Assignee: bugs at gluster.org
Reporter: rkavunga at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
There are quite a few memory leaks identified for client side xlators.
1) xlators/cluster/afr/src/afr.c ---> this->local_pool is not freed
2) xlators/cluster/ec/src/ec.c ----> this->itable is not freed
3) protocol/client/src/client.c ----> this->local_pool is not freed
I will add more to this list in case if I found any other leaks
Version-Release number of selected component (if applicable):
How reproducible:
Steps to Reproduce:
1.
2.
3.
Actual results:
Expected results:
Additional info:
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 00:19:01 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 00:19:01 +0000
Subject: [Bugs] [Bug 1716695] Fix memory leaks that are present even after
an xlator fini [client side xlator]
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716695
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22806
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 00:19:02 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 00:19:02 +0000
Subject: [Bugs] [Bug 1716695] Fix memory leaks that are present even after
an xlator fini [client side xlator]
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716695
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22806 (afr/fini: Free local_pool data during
an afr fini) posted (#1) for review on master by mohammed rafi kc
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 00:20:14 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 00:20:14 +0000
Subject: [Bugs] [Bug 1716695] Fix memory leaks that are present even after
an xlator fini [client side xlator]
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716695
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22807
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 00:20:15 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 00:20:15 +0000
Subject: [Bugs] [Bug 1716695] Fix memory leaks that are present even after
an xlator fini [client side xlator]
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716695
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22807 (ec/fini: Free itable during an ec
fini) posted (#1) for review on master by mohammed rafi kc
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 02:59:31 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 02:59:31 +0000
Subject: [Bugs] [Bug 789278] Issues reported by Coverity static analysis tool
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=789278
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22809
--
You are receiving this mail because:
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 02:59:33 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 02:59:33 +0000
Subject: [Bugs] [Bug 789278] Issues reported by Coverity static analysis tool
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=789278
--- Comment #1633 from Worker Ant ---
REVIEW: https://review.gluster.org/22809 (posix: coverity fix) posted (#1) for
review on master by MOHIT AGRAWAL
--
You are receiving this mail because:
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 04:17:20 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 04:17:20 +0000
Subject: [Bugs] [Bug 1716626] Invalid memory access while executing
cleanup_and_exit
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716626
Vivek Das changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |vdas at redhat.com
Blocks| |1696809
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 04:17:24 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 04:17:24 +0000
Subject: [Bugs] [Bug 1716626] Invalid memory access while executing
cleanup_and_exit
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716626
RHEL Product and Program Management changed:
What |Removed |Added
----------------------------------------------------------------------------
Rule Engine Rule| |Gluster: Auto pm_ack for
| |dev&qe approved in-flight
| |RHGS3.5 BZs
Rule Engine Rule| |665
Target Release|--- |RHGS 3.5.0
Rule Engine Rule| |666
Rule Engine Rule| |327
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:07:27 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:07:27 +0000
Subject: [Bugs] [Bug 1716760] New: Make debugging hung frames easier
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716760
Bug ID: 1716760
Summary: Make debugging hung frames easier
Product: Red Hat Gluster Storage
Version: rhgs-3.5
Status: NEW
Component: core
Assignee: atumball at redhat.com
Reporter: pkarampu at redhat.com
QA Contact: rhinduja at redhat.com
CC: bugs at gluster.org, rhs-bugs at redhat.com,
sankarshan at redhat.com, storage-qa-internal at redhat.com
Depends On: 1714098
Target Milestone: ---
Classification: Red Hat
+++ This bug was initially created as a clone of Bug #1714098 +++
Description of problem:
At the moment new stack doesn't populate frame->root->unique in all cases. This
makes it difficult to debug hung frames by examining successive state dumps.
Fuse and server xlator populate it whenever they can, but other xlators won't
be able to assign one when they need to create a new frame/stack. What we need
is for unique to be correct. If a stack with same unique is present in
successive statedumps, that means the same operation is still in progress. This
makes finding hung frames part of debugging hung frames easier.
Version-Release number of selected component (if applicable):
How reproducible:
Steps to Reproduce:
1.
2.
3.
Actual results:
Expected results:
Additional info:
--- Additional comment from Worker Ant on 2019-05-27 06:27:36 UTC ---
REVIEW: https://review.gluster.org/22773 (stack: Make sure to have unique
call-stacks in all cases) posted (#1) for review on master by Pranith Kumar
Karampuri
--- Additional comment from Worker Ant on 2019-05-30 15:55:06 UTC ---
REVIEW: https://review.gluster.org/22773 (stack: Make sure to have unique
call-stacks in all cases) merged (#4) on master by Amar Tumballi
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1714098
[Bug 1714098] Make debugging hung frames easier
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:07:27 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:07:27 +0000
Subject: [Bugs] [Bug 1714098] Make debugging hung frames easier
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714098
Pranith Kumar K changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1716760
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1716760
[Bug 1716760] Make debugging hung frames easier
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 05:07:30 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:07:30 +0000
Subject: [Bugs] [Bug 1716760] Make debugging hung frames easier
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716760
RHEL Product and Program Management changed:
What |Removed |Added
----------------------------------------------------------------------------
Rule Engine Rule| |Gluster: set proposed
| |release flag for new BZs at
| |RHGS
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:08:13 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:08:13 +0000
Subject: [Bugs] [Bug 1716760] Make debugging hung frames easier
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716760
Pranith Kumar K changed:
What |Removed |Added
----------------------------------------------------------------------------
Assignee|atumball at redhat.com |pkarampu at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:18:20 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:18:20 +0000
Subject: [Bugs] [Bug 1716766] New: [Thin-arbiter] TA process is not picking
24007 as port while starting up
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716766
Bug ID: 1716766
Summary: [Thin-arbiter] TA process is not picking 24007 as port
while starting up
Product: GlusterFS
Version: mainline
Status: NEW
Component: replicate
Assignee: bugs at gluster.org
Reporter: aspandey at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
TA process is not picking 24007 as port while starting up.
Problem:
In unit file of TA process we have been using ta-vol as volume id and also
ta-vol-server.transport.socket.listen-port=24007
In our volume file for TA process we only consider volname as "ta" and not as
"ta-vol".
That's why it was not able to assign this port number to or ta process as in
volume file it will try to find server xlato as ta-vol
volume ta-server <<<<<<<<< not ta-vol
46 type protocol/server
47 option transport.listen-backlog 10
48 option transport.socket.keepalive-count 9
49 option transport.socket.keepalive-interval 2
50 option transport.socket.keepalive-time 20
51 option transport.tcp-user-timeout 0
52 option transport.socket.keepalive 1
53 option auth.addr./mnt/thin-arbiter.allow *
54 option auth-path /mnt/thin-arbiter
55 option transport.address-family inet
56 option transport-type tcp
57 subvolumes ta-io-stats
58 end-volume
Solution:
Just need to change the command which Unit file is going to execute.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 05:25:41 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:25:41 +0000
Subject: [Bugs] [Bug 1716760] Make debugging hung frames easier
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716760
Pranith Kumar K changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #2 from Pranith Kumar K ---
Patch link: https://code.engineering.redhat.com/gerrit/#/c/172304
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:30:09 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:30:09 +0000
Subject: [Bugs] [Bug 1716440] SMBD thread panics when connected to from OS X
machine
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716440
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags| |needinfo?(ryan at magenta.tv)
--- Comment #1 from Anoop C S ---
(In reply to ryan from comment #0)
> [VFS]
> vfs objects = glusterfs
'fruit' and 'stream_xattr' vfs modules are recommended to be loaded while
connecting/accessing/operating on SMB shares using Samba from Mac OS X clients.
Can you re-try connecting to shares with following additional settings:
vfs objects = fruit streams_xattr glusterfs
fruit:encoding = native
Also please add the following in [global] section:
ea support = yes
fruit:aapl = yes
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 05:39:07 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:39:07 +0000
Subject: [Bugs] [Bug 1716626] Invalid memory access while executing
cleanup_and_exit
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716626
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |MODIFIED
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:41:55 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:41:55 +0000
Subject: [Bugs] [Bug 1714536] geo-rep: With heavy rename workload geo-rep
log if flooded
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714536
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
Version|unspecified |rhgs-3.5
Assignee|sunkumar at redhat.com |khiremat at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:47:38 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:47:38 +0000
Subject: [Bugs] [Bug 1703948] Self-heal daemon resources are not cleaned
properly after a ec fini
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703948
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22810
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:47:39 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:47:39 +0000
Subject: [Bugs] [Bug 1703948] Self-heal daemon resources are not cleaned
properly after a ec fini
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703948
--- Comment #6 from Worker Ant ---
REVIEW: https://review.gluster.org/22810 (xlator/log: Add more logging in
xlator_is_cleanup_starting) posted (#1) for review on master by mohammed rafi
kc
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:57:14 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:57:14 +0000
Subject: [Bugs] [Bug 1714536] geo-rep: With heavy rename workload geo-rep
log if flooded
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714536
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 05:57:15 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 05:57:15 +0000
Subject: [Bugs] [Bug 789278] Issues reported by Coverity static analysis tool
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=789278
--- Comment #1634 from Worker Ant ---
REVIEW: https://review.gluster.org/22801 (glusterd: coverity fix) merged (#4)
on master by Atin Mukherjee
--
You are receiving this mail because:
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:01:28 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:01:28 +0000
Subject: [Bugs] [Bug 1716766] [Thin-arbiter] TA process is not picking 24007
as port while starting up
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716766
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22811
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:01:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:01:29 +0000
Subject: [Bugs] [Bug 1716766] [Thin-arbiter] TA process is not picking 24007
as port while starting up
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716766
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22811 (cluster/replicate: Modify command in
unit file to assign port correctly) posted (#1) for review on master by Ashish
Pandey
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:04:55 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:04:55 +0000
Subject: [Bugs] [Bug 1716766] [Thin-arbiter] TA process is not picking 24007
as port while starting up
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716766
Ashish Pandey changed:
What |Removed |Added
----------------------------------------------------------------------------
Assignee|bugs at gluster.org |aspandey at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:15:05 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:15:05 +0000
Subject: [Bugs] [Bug 1716790] New: geo-rep: Rename with same name testcase
is failing with EV Volume
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716790
Bug ID: 1716790
Summary: geo-rep: Rename with same name testcase is failing
with EV Volume
Product: GlusterFS
Version: mainline
Status: NEW
Component: geo-replication
Assignee: bugs at gluster.org
Reporter: khiremat at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
Rename with same name testcase is failing with EC Volume
Version-Release number of selected component (if applicable):
mainline
How reproducible:
Occasionally
Steps to Reproduce:
Occasional upstream regression run failures
1. https://build.gluster.org/job/centos7-regression/6281/console
2. https://build.gluster.org/job/centos7-regression/6278/
Actual results:
geo-rep EC volume rename testcase failed occasionally
Expected results:
geo-rep EC volume rename testcase should always pass
Additional info:
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:16:09 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:16:09 +0000
Subject: [Bugs] [Bug 1716790] geo-rep: Rename with same name testcase is
failing with EV Volume
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716790
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Assignee|bugs at gluster.org |sacharya at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:22:20 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:22:20 +0000
Subject: [Bugs] [Bug 1716790] geo-rep: Rename with same destination name
test case occasionally fails on EC Volume
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716790
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Summary|geo-rep: Rename with same |geo-rep: Rename with same
|name testcase is failing |destination name test case
|with EV Volume |occasionally fails on EC
| |Volume
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 06:25:25 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:25:25 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #678 from Worker Ant ---
REVIEW: https://review.gluster.org/22804 (tests/geo-rep: Remove a rename test
case on EC volume) merged (#5) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:25:49 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:25:49 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
--- Comment #53 from Worker Ant ---
REVIEW: https://review.gluster.org/22803 (tests/geo-rep: Add geo-rep glusterd
test cases) merged (#2) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:55:25 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:55:25 +0000
Subject: [Bugs] [Bug 1716440] SMBD thread panics when connected to from OS X
machine
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716440
ryan at magenta.tv changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags|needinfo?(ryan at magenta.tv) |
--- Comment #2 from ryan at magenta.tv ---
Hi Anoop,
Thanks for getting back to me.
I've tried your suggestion but unfortunately the issue still remains. Here is
my updated smb.conf:
[global]
security = user
netbios name = NAS01
clustering = no
server signing = no
max log size = 10000
log file = /var/log/samba/log-%M-test.smbd
logging = file
log level = 10
passdb backend = tdbsam
guest account = nobody
map to guest = bad user
force directory mode = 0777
force create mode = 0777
create mask = 0777
directory mask = 0777
store dos attributes = yes
load printers = no
printing = bsd
printcap name = /dev/null
disable spoolss = yes
glusterfs:volfile_server = localhost
ea support = yes
fruit:aapl = yes
kernel share modes = No
[VFS]
vfs objects = fruit streams_xattr glusterfs
fruit:encoding = native
glusterfs:volume = mcv02
path = /
read only = no
guest ok = yes
This time when creating a new folder at the root of the share, it creates, then
disappears, sometimes coming back, sometimes not.
When I was able to traverse into a sub-folder, the same error is received.
I will attach the debug level 10 logs to the bug.
Many thanks for you help,
Ryan
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 06:56:23 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 06:56:23 +0000
Subject: [Bugs] [Bug 1716440] SMBD thread panics when connected to from OS X
machine
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716440
--- Comment #3 from ryan at magenta.tv ---
Created attachment 1576920
--> https://bugzilla.redhat.com/attachment.cgi?id=1576920&action=edit
Debug level 10 log of issue after adding streams_xattr and fruit
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 07:30:49 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 07:30:49 +0000
Subject: [Bugs] [Bug 1705884] Image size as reported from the fuse mount is
incorrect
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1705884
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-04 07:30:49
--- Comment #4 from Worker Ant ---
REVIEW: https://review.gluster.org/22681 (features/shard: Fix block-count
accounting upon truncate to lower size) merged (#6) on master by Xavi Hernandez
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 07:52:34 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 07:52:34 +0000
Subject: [Bugs] [Bug 1716812] New: Failed to create volume which
transport_type is "tcp, rdma"
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716812
Bug ID: 1716812
Summary: Failed to create volume which transport_type is
"tcp,rdma"
Product: GlusterFS
Version: 4.1
Hardware: x86_64
OS: Linux
Status: NEW
Component: glusterd
Severity: high
Assignee: bugs at gluster.org
Reporter: guol-fnst at cn.fujitsu.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
gluster volume create 11 transport tcp,rdma 193.168.141.101:/tmp/11
193.168.141.101:/tmp/12 force
volume create: 11: failed: Failed to create volume files
Version-Release number of selected component (if applicable):
# gluster --version
glusterfs 4.1.8
Repository revision: git://git.gluster.org/glusterfs.git
Copyright (c) 2006-2016 Red Hat, Inc.
GlusterFS comes with ABSOLUTELY NO WARRANTY.
It is licensed to you under your choice of the GNU Lesser
General Public License, version 3 or any later version (LGPLv3
or later), or the GNU General Public License, version 2 (GPLv2),
in all cases as published by the Free Software Foundation.
# ip a
1: lo: mtu 65536 qdisc noqueue state UNKNOWN qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: ens192: mtu 1500 qdisc pfifo_fast state UP
qlen 1000
link/ether 00:50:56:9c:8b:a9 brd ff:ff:ff:ff:ff:ff
inet 193.168.141.101/16 brd 193.168.255.255 scope global dynamic ens192
valid_lft 2591093sec preferred_lft 2591093sec
inet6 fe80::250:56ff:fe9c:8ba9/64 scope link
valid_lft forever preferred_lft forever
3: ens224: mtu 1500 qdisc pfifo_fast state UP
qlen 1000
link/ether 00:50:56:9c:53:58 brd ff:ff:ff:ff:ff:ff
How reproducible:
Steps to Reproduce:
1.rxe_cfg start
2.rxe_cfg add ens192
3.gluster volume create 11 transport tcp,rdma 193.168.141.101:/tmp/11
193.168.141.101:/tmp/12 force
Actual results:
volume create: 11: failed: Failed to create volume files
Expected results:
Success to create volume
Additional info:
[2019-06-04 07:36:45.966125] I [MSGID: 100030] [glusterfsd.c:2741:main]
0-glusterd: Started running glusterd version 4.1.8 (args: glusterd
--xlator-option *.upgrade=on -N)
[2019-06-04 07:36:45.970884] I [MSGID: 106478] [glusterd.c:1423:init]
0-management: Maximum allowed open file descriptors set to 65536
[2019-06-04 07:36:45.970900] I [MSGID: 106479] [glusterd.c:1481:init]
0-management: Using /var/lib/glusterd as working directory
[2019-06-04 07:36:45.970906] I [MSGID: 106479] [glusterd.c:1486:init]
0-management: Using /var/run/gluster as pid file working directory
[2019-06-04 07:36:45.973455] E [rpc-transport.c:284:rpc_transport_load]
0-rpc-transport: /usr/lib64/glusterfs/4.1.8/rpc-transport/rdma.so: cannot open
shared object file: No such file or directory
[2019-06-04 07:36:45.973468] W [rpc-transport.c:288:rpc_transport_load]
0-rpc-transport: volume 'rdma.management': transport-type 'rdma' is not valid
or not found on this machine
[2019-06-04 07:36:45.973473] W [rpcsvc.c:1781:rpcsvc_create_listener]
0-rpc-service: cannot create listener, initing the transport failed
[2019-06-04 07:36:45.973478] E [MSGID: 106244] [glusterd.c:1764:init]
0-management: creation of 1 listeners failed, continuing with succeeded
transport
[2019-06-04 07:36:45.976348] I [MSGID: 106513]
[glusterd-store.c:2240:glusterd_restore_op_version] 0-glusterd: retrieved
op-version: 31202
[2019-06-04 07:36:45.977372] I [MSGID: 106544]
[glusterd.c:158:glusterd_uuid_init] 0-management: retrieved UUID:
79e7e129-d041-48b6-b1d0-746c55d148fc
[2019-06-04 07:36:45.989706] I [MSGID: 106194]
[glusterd-store.c:3850:glusterd_store_retrieve_missed_snaps_list] 0-management:
No missed snaps list.
Final graph:
+------------------------------------------------------------------------------+
1: volume management
2: type mgmt/glusterd
3: option rpc-auth.auth-glusterfs on
4: option rpc-auth.auth-unix on
5: option rpc-auth.auth-null on
6: option rpc-auth-allow-insecure on
7: option transport.listen-backlog 10
8: option upgrade on
9: option event-threads 1
10: option ping-timeout 0
11: option transport.socket.read-fail-log off
12: option transport.socket.keepalive-interval 2
13: option transport.socket.keepalive-time 10
14: option transport-type rdma
15: option working-directory /var/lib/glusterd
16: end-volume
17:
+------------------------------------------------------------------------------+
[2019-06-04 07:36:46.005401] I [MSGID: 101190]
[event-epoll.c:617:event_dispatch_epoll_worker] 0-epoll: Started thread with
index 1
[2019-06-04 07:36:46.006879] W [glusterfsd.c:1514:cleanup_and_exit]
(-->/usr/lib64/libpthread.so.0(+0x7dd5) [0x7f55547bbdd5]
-->glusterd(glusterfs_sigwaiter+0xe5) [0x55c659e7dd65]
-->glusterd(cleanup_and_exit+0x6b) [0x55c659e7db8b] ) 0-: received signum (15),
shutting down
[2019-06-04 07:36:46.006997] E
[rpcsvc.c:1536:rpcsvc_program_unregister_portmap] 0-rpc-service: Could not
unregister with portmap
[2019-06-04 07:36:46.007004] E [rpcsvc.c:1662:rpcsvc_program_unregister]
0-rpc-service: portmap unregistration of program failed
[2019-06-04 07:36:46.007008] E [rpcsvc.c:1708:rpcsvc_program_unregister]
0-rpc-service: Program unregistration failed: GlusterD svc cli, Num: 1238463,
Ver: 2, Port: 0
[2019-06-04 07:36:46.007061] E
[rpcsvc.c:1536:rpcsvc_program_unregister_portmap] 0-rpc-service: Could not
unregister with portmap
[2019-06-04 07:36:46.007066] E [rpcsvc.c:1662:rpcsvc_program_unregister]
0-rpc-service: portmap unregistration of program failed
[2019-06-04 07:36:46.007070] E [rpcsvc.c:1708:rpcsvc_program_unregister]
0-rpc-service: Program unregistration failed: Gluster Handshake, Num: 14398633,
Ver: 2, Port: 0
[2019-06-04 07:37:18.784525] I [MSGID: 100030] [glusterfsd.c:2741:main]
0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 4.1.8 (args:
/usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO)
[2019-06-04 07:37:18.787926] I [MSGID: 106478] [glusterd.c:1423:init]
0-management: Maximum allowed open file descriptors set to 65536
[2019-06-04 07:37:18.787944] I [MSGID: 106479] [glusterd.c:1481:init]
0-management: Using /var/lib/glusterd as working directory
[2019-06-04 07:37:18.787950] I [MSGID: 106479] [glusterd.c:1486:init]
0-management: Using /var/run/gluster as pid file working directory
[2019-06-04 07:37:18.814752] W [MSGID: 103071]
[rdma.c:4629:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm event channel
creation failed [No such device]
[2019-06-04 07:37:18.814780] W [MSGID: 103055] [rdma.c:4938:init]
0-rdma.management: Failed to initialize IB Device
[2019-06-04 07:37:18.814786] W [rpc-transport.c:351:rpc_transport_load]
0-rpc-transport: 'rdma' initialization failed
[2019-06-04 07:37:18.814844] W [rpcsvc.c:1781:rpcsvc_create_listener]
0-rpc-service: cannot create listener, initing the transport failed
[2019-06-04 07:37:18.814852] E [MSGID: 106244] [glusterd.c:1764:init]
0-management: creation of 1 listeners failed, continuing with succeeded
transport
[2019-06-04 07:37:19.617049] I [MSGID: 106513]
[glusterd-store.c:2240:glusterd_restore_op_version] 0-glusterd: retrieved
op-version: 31202
[2019-06-04 07:37:19.617342] I [MSGID: 106544]
[glusterd.c:158:glusterd_uuid_init] 0-management: retrieved UUID:
79e7e129-d041-48b6-b1d0-746c55d148fc
[2019-06-04 07:37:19.626546] I [MSGID: 106194]
[glusterd-store.c:3850:glusterd_store_retrieve_missed_snaps_list] 0-management:
No missed snaps list.
Final graph:
+------------------------------------------------------------------------------+
1: volume management
2: type mgmt/glusterd
3: option rpc-auth.auth-glusterfs on
4: option rpc-auth.auth-unix on
5: option rpc-auth.auth-null on
6: option rpc-auth-allow-insecure on
7: option transport.listen-backlog 10
8: option event-threads 1
9: option ping-timeout 0
10: option transport.socket.read-fail-log off
11: option transport.socket.keepalive-interval 2
12: option transport.socket.keepalive-time 10
13: option transport-type rdma
14: option working-directory /var/lib/glusterd
15: end-volume
16:
+------------------------------------------------------------------------------+
[2019-06-04 07:37:19.626791] I [MSGID: 101190]
[event-epoll.c:617:event_dispatch_epoll_worker] 0-epoll: Started thread with
index 1
[2019-06-04 07:37:20.874611] W [MSGID: 101095]
[xlator.c:181:xlator_volopt_dynload] 0-xlator:
/usr/lib64/glusterfs/4.1.8/xlator/nfs/server.so: cannot open shared object
file: No such file or directory
[2019-06-04 07:37:20.889571] E [MSGID: 106068]
[glusterd-volgen.c:1034:volgen_write_volfile] 0-management: failed to create
volfile
[2019-06-04 07:37:20.889588] E
[glusterd-volgen.c:6727:glusterd_create_volfiles] 0-management: Could not
generate gfproxy client volfiles
[2019-06-04 07:37:20.889601] E [MSGID: 106122]
[glusterd-syncop.c:1482:gd_commit_op_phase] 0-management: Commit of operation
'Volume Create' failed on localhost : Failed to create volume files
[2019-06-04 07:38:49.194175] W [MSGID: 101095]
[xlator.c:181:xlator_volopt_dynload] 0-xlator:
/usr/lib64/glusterfs/4.1.8/xlator/nfs/server.so: cannot open shared object
file: No such file or directory
[2019-06-04 07:38:49.211380] E [MSGID: 106068]
[glusterd-volgen.c:1034:volgen_write_volfile] 0-management: failed to create
volfile
[2019-06-04 07:38:49.211407] E
[glusterd-volgen.c:6727:glusterd_create_volfiles] 0-management: Could not
generate gfproxy client volfiles
[2019-06-04 07:38:49.211433] E [MSGID: 106122]
[glusterd-syncop.c:1482:gd_commit_op_phase] 0-management: Commit of operation
'Volume Create' failed on localhost : Failed to create volume files
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 07:58:44 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 07:58:44 +0000
Subject: [Bugs] [Bug 1716812] Failed to create volume which transport_type
is "tcp, rdma"
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716812
--- Comment #1 from guolei ---
Test is ok on glusterfs3.12.9 ,failed on glusterfs3.13.2 and later version.
generate_client_volfiles (glusterd_volinfo_t *volinfo,
glusterd_client_type_t client_type)
{
int i = 0;
int ret = -1;
char filepath[PATH_MAX] = {0,};
char *types[] = {NULL, NULL, NULL};
dict_t *dict = NULL;
xlator_t *this = NULL;
gf_transport_type type = GF_TRANSPORT_TCP;
this = THIS;
enumerate_transport_reqs (volinfo->transport_type, types);
dict = dict_new ();
if (!dict)
goto out;
for (i = 0; types[i]; i++) {
memset (filepath, 0, sizeof (filepath));
ret = dict_set_str (dict, "client-transport-type", types[i]);
if (ret)
goto out;
type = transport_str_to_type (types[i]);
ret = dict_set_uint32 (dict, "trusted-client", client_type);
if (ret)
goto out;
if (client_type == GF_CLIENT_TRUSTED) {
ret = glusterd_get_trusted_client_filepath (filepath,
volinfo,
type);
} else if (client_type == GF_CLIENT_TRUSTED_PROXY) {
glusterd_get_gfproxy_client_volfile (volinfo,
filepath,
PATH_MAX);
<---------------------------- Maybe this is the problem? transport type should
be passed to glusterd_get_gfproxy_client_volfile .Or filepath is NULL.
ret = dict_set_str (dict, "gfproxy-client", "on");
} else {
ret = glusterd_get_client_filepath (filepath,
volinfo,
type);
}
if (ret) {
gf_msg (this->name, GF_LOG_ERROR, EINVAL,
GD_MSG_INVALID_ENTRY,
"Received invalid transport-type");
goto out;
}
* ret = generate_single_transport_client_volfile (volinfo,
filepath,
dict);*
if (ret)
goto out;
}
/* Generate volfile for rebalance process */
glusterd_get_rebalance_volfile (volinfo, filepath, PATH_MAX);
ret = build_rebalance_volfile (volinfo, filepath, dict);
if (ret) {
gf_msg (this->name, GF_LOG_ERROR, 0,
GD_MSG_VOLFILE_CREATE_FAIL,
"Failed to create rebalance volfile for %s",
volinfo->volname);
goto out;
}
out:
if (dict)
dict_unref (dict);
gf_msg_trace ("glusterd", 0, "Returning %d", ret);
return ret;
}
void
glusterd_get_gfproxy_client_volfile (glusterd_volinfo_t *volinfo,
char *path, int path_len)
{
char workdir[PATH_MAX] = {0, };
glusterd_conf_t *priv = THIS->private;
GLUSTERD_GET_VOLUME_DIR (workdir, volinfo, priv);
switch (volinfo->transport_type) {
case GF_TRANSPORT_TCP:
snprintf (path, path_len,
"%s/trusted-%s.tcp-gfproxy-fuse.vol",
workdir, volinfo->volname);
break;
case GF_TRANSPORT_RDMA:
snprintf (path, path_len,
"%s/trusted-%s.rdma-gfproxy-fuse.vol",
workdir, volinfo->volname);
break;
default:
break;
}
}
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 09:05:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:05:10 +0000
Subject: [Bugs] [Bug 1716830] New: DHT: directory permissions are wiped out
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
Bug ID: 1716830
Summary: DHT: directory permissions are wiped out
Product: GlusterFS
Version: mainline
Status: NEW
Component: distribute
Assignee: bugs at gluster.org
Reporter: nbalacha at redhat.com
CC: bugs at gluster.org, rhs-bugs at redhat.com,
sankarshan at redhat.com, saraut at redhat.com,
storage-qa-internal at redhat.com
Depends On: 1716821
Target Milestone: ---
Classification: Community
+++ This bug was initially created as a clone of Bug #1716821 +++
Description of problem:
A sequence of steps can wipe out the permissions set on a directory.
Version-Release number of selected component (if applicable):
RHGS 3.5.0
How reproducible:
Consistently
Steps to Reproduce:
[root at rhgs313-6 ~]# gluster volume create vol1
192.168.122.6:/bricks/brick1/vol1-1
volume create: vol1: success: please start the volume to access data
[root at rhgs313-6 ~]# gluster v start vol1
volume start: vol1: success
[root at rhgs313-6 ~]# mount -t glusterfs -s 192.168.122.6:/vol1 /mnt/fuse1
[root at rhgs313-6 fuse]# cd /mnt/fuse1
[root at rhgs313-6 fuse1]# mkdir dir1
[root at rhgs313-6 fuse1]# cd dir1/
[root at rhgs313-6 dir1]# getx /bricks/brick1/vol1-*/dir1
getfattr: Removing leading '/' from absolute path names
# file: bricks/brick1/vol1-1/dir1
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a756e6c6162656c65645f743a733000
trusted.gfid=0xbf9444c0f8614d81a5758ed801e9f7e0
trusted.glusterfs.dht=0x000000000000000000000000ffffffff
trusted.glusterfs.mdata=0x010000000000000000000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6
[root at rhgs313-6 dir1]# gluster v add-brick vol1
192.168.122.6:/bricks/brick1/vol1-2 force
volume add-brick: success
[root at rhgs313-6 dir1]# ll
total 0
Check the directory permissions and xattrs on the bricks
[root at rhgs313-6 dir1]# ll /bricks/brick1/vol1-*
/bricks/brick1/vol1-1:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
/bricks/brick1/vol1-2:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
[root at rhgs313-6 dir1]# getx /bricks/brick1/vol1-*/dir1
getfattr: Removing leading '/' from absolute path names
# file: bricks/brick1/vol1-1/dir1
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a756e6c6162656c65645f743a733000
trusted.gfid=0xbf9444c0f8614d81a5758ed801e9f7e0
trusted.glusterfs.dht=0x000000000000000000000000ffffffff
trusted.glusterfs.mdata=0x010000000000000000000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6
# file: bricks/brick1/vol1-2/dir1
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a756e6c6162656c65645f743a733000
trusted.gfid=0xbf9444c0f8614d81a5758ed801e9f7e0
trusted.glusterfs.dht=0x00000000000000000000000000000000
trusted.glusterfs.mdata=0x010000000000000000000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6
>From the mount point, cd one level up and then back again into dir1.
[root at rhgs313-6 dir1]# cd ..
[root at rhgs313-6 fuse1]# cd dir1
[root at rhgs313-6 dir1]# ll /bricks/brick1/vol1-*
Actual results:
[root at rhgs313-6 dir1]# ll /bricks/brick1/vol1-*
/bricks/brick1/vol1-1:
total 0
d---------. 2 root root 6 Jun 4 13:50 dir1
/bricks/brick1/vol1-2:
total 0
d---------. 2 root root 6 Jun 4 13:50 dir1
Expected results:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
/bricks/brick1/vol1-2:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
Additional info:
--- Additional comment from RHEL Product and Program Management on 2019-06-04
08:29:48 UTC ---
This bug is automatically being proposed for the next minor release of Red Hat
Gluster Storage by setting the release flag 'rhgs?3.5.0' to '?'.
If this bug should be proposed for a different release, please manually change
the proposed release flag.
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1716821
[Bug 1716821] DHT: directory permissions are wiped out
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 09:23:40 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:23:40 +0000
Subject: [Bugs] [Bug 1716830] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
Nithya Balachandran changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
Assignee|bugs at gluster.org |nbalacha at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 09:28:43 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:28:43 +0000
Subject: [Bugs] [Bug 1716830] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22813
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 09:28:44 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:28:44 +0000
Subject: [Bugs] [Bug 1716830] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22813 (cluster/dht: Fix directory perms
during selfheal) posted (#1) for review on master by N Balachandran
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 09:46:17 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:46:17 +0000
Subject: [Bugs] [Bug 1716848] New: DHT: directory permissions are wiped out
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716848
Bug ID: 1716848
Summary: DHT: directory permissions are wiped out
Product: GlusterFS
Version: 6
Status: NEW
Component: distribute
Assignee: bugs at gluster.org
Reporter: nbalacha at redhat.com
CC: bugs at gluster.org, rhs-bugs at redhat.com,
sankarshan at redhat.com, saraut at redhat.com,
storage-qa-internal at redhat.com
Depends On: 1716821, 1716830
Target Milestone: ---
Classification: Community
+++ This bug was initially created as a clone of Bug #1716830 +++
+++ This bug was initially created as a clone of Bug #1716821 +++
Description of problem:
A sequence of steps can wipe out the permissions set on a directory.
Version-Release number of selected component (if applicable):
RHGS 3.5.0
How reproducible:
Consistently
Steps to Reproduce:
[root at rhgs313-6 ~]# gluster volume create vol1
192.168.122.6:/bricks/brick1/vol1-1
volume create: vol1: success: please start the volume to access data
[root at rhgs313-6 ~]# gluster v start vol1
volume start: vol1: success
[root at rhgs313-6 ~]# mount -t glusterfs -s 192.168.122.6:/vol1 /mnt/fuse1
[root at rhgs313-6 fuse]# cd /mnt/fuse1
[root at rhgs313-6 fuse1]# mkdir dir1
[root at rhgs313-6 fuse1]# cd dir1/
[root at rhgs313-6 dir1]# getx /bricks/brick1/vol1-*/dir1
getfattr: Removing leading '/' from absolute path names
# file: bricks/brick1/vol1-1/dir1
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a756e6c6162656c65645f743a733000
trusted.gfid=0xbf9444c0f8614d81a5758ed801e9f7e0
trusted.glusterfs.dht=0x000000000000000000000000ffffffff
trusted.glusterfs.mdata=0x010000000000000000000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6
[root at rhgs313-6 dir1]# gluster v add-brick vol1
192.168.122.6:/bricks/brick1/vol1-2 force
volume add-brick: success
[root at rhgs313-6 dir1]# ll
total 0
Check the directory permissions and xattrs on the bricks
[root at rhgs313-6 dir1]# ll /bricks/brick1/vol1-*
/bricks/brick1/vol1-1:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
/bricks/brick1/vol1-2:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
[root at rhgs313-6 dir1]# getx /bricks/brick1/vol1-*/dir1
getfattr: Removing leading '/' from absolute path names
# file: bricks/brick1/vol1-1/dir1
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a756e6c6162656c65645f743a733000
trusted.gfid=0xbf9444c0f8614d81a5758ed801e9f7e0
trusted.glusterfs.dht=0x000000000000000000000000ffffffff
trusted.glusterfs.mdata=0x010000000000000000000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6
# file: bricks/brick1/vol1-2/dir1
security.selinux=0x756e636f6e66696e65645f753a6f626a6563745f723a756e6c6162656c65645f743a733000
trusted.gfid=0xbf9444c0f8614d81a5758ed801e9f7e0
trusted.glusterfs.dht=0x00000000000000000000000000000000
trusted.glusterfs.mdata=0x010000000000000000000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6000000005cf629cf00000000302dadf6
>From the mount point, cd one level up and then back again into dir1.
[root at rhgs313-6 dir1]# cd ..
[root at rhgs313-6 fuse1]# cd dir1
[root at rhgs313-6 dir1]# ll /bricks/brick1/vol1-*
Actual results:
[root at rhgs313-6 dir1]# ll /bricks/brick1/vol1-*
/bricks/brick1/vol1-1:
total 0
d---------. 2 root root 6 Jun 4 13:50 dir1
/bricks/brick1/vol1-2:
total 0
d---------. 2 root root 6 Jun 4 13:50 dir1
Expected results:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
/bricks/brick1/vol1-2:
total 0
drwxr-xr-x. 2 root root 6 Jun 4 13:50 dir1
Additional info:
--- Additional comment from RHEL Product and Program Management on 2019-06-04
08:29:48 UTC ---
This bug is automatically being proposed for the next minor release of Red Hat
Gluster Storage by setting the release flag 'rhgs?3.5.0' to '?'.
If this bug should be proposed for a different release, please manually change
the proposed release flag.
--- Additional comment from Worker Ant on 2019-06-04 09:28:44 UTC ---
REVIEW: https://review.gluster.org/22813 (cluster/dht: Fix directory perms
during selfheal) posted (#1) for review on master by N Balachandran
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1716821
[Bug 1716821] DHT: directory permissions are wiped out
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
[Bug 1716830] DHT: directory permissions are wiped out
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 09:46:17 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:46:17 +0000
Subject: [Bugs] [Bug 1716830] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
Nithya Balachandran changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1716848
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1716848
[Bug 1716848] DHT: directory permissions are wiped out
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 09:47:12 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:47:12 +0000
Subject: [Bugs] [Bug 1716848] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716848
Nithya Balachandran changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
Assignee|bugs at gluster.org |nbalacha at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 09:51:37 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:51:37 +0000
Subject: [Bugs] [Bug 1716848] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716848
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22814
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 09:51:38 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:51:38 +0000
Subject: [Bugs] [Bug 1716848] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716848
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22814 (cluster/dht: Fix directory perms
during selfheal) posted (#1) for review on release-6 by N Balachandran
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Tue Jun 4 09:53:36 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 09:53:36 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
--- Comment #54 from Worker Ant ---
REVIEW: https://review.gluster.org/22799 (lcov: run more fops on translators)
merged (#3) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:28:42 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:28:42 +0000
Subject: [Bugs] [Bug 1716870] New: Came up with a script to analyze strace
outputs from bricks
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716870
Bug ID: 1716870
Summary: Came up with a script to analyze strace outputs from
bricks
Product: GlusterFS
Version: mainline
Status: NEW
Component: scripts
Assignee: bugs at gluster.org
Reporter: pkarampu at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Debugging performance issues often includes comparison of brick strace files to
compare number of syscalls/maximum-latencies per syscall etc with previous
runs.
This script helps in getting these numbers.
Running it creates 3 types of files:
1) syscalls-summary.txt - Prints per syscall counts
2) -latency.txt - This is an intermediate file where all 'syscall'
calls from all the strace files will be listed.
3) per-fop-latency.txt - Per syscall it prints top maximum latencies
observed.
Assumes the files in strace-dir are created using the following command:
$ strace -ff -T -p -o
Sample output of syscalls-summary.txt:
49857 chmod
49906 chown
97542 close
650309 fgetxattr
18 flistxattr
....
Sample output of per-fop-latency.txt:
--chmod--
0.000216
0.000254
0.000266
...
--unlink--
0.020208
0.025084
0.027231
...
Version-Release number of selected component (if applicable):
How reproducible:
Steps to Reproduce:
1.
2.
3.
Actual results:
Expected results:
Additional info:
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:29:12 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:29:12 +0000
Subject: [Bugs] [Bug 1716871] New: Image size as reported from the fuse
mount is incorrect
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716871
Bug ID: 1716871
Summary: Image size as reported from the fuse mount is
incorrect
Product: GlusterFS
Version: 6
Hardware: x86_64
OS: Linux
Status: NEW
Component: sharding
Severity: high
Assignee: bugs at gluster.org
Reporter: kdhananj at redhat.com
QA Contact: bugs at gluster.org
CC: bugs at gluster.org, kdhananj at redhat.com, pasik at iki.fi,
rhs-bugs at redhat.com, sabose at redhat.com,
sankarshan at redhat.com, sasundar at redhat.com,
storage-qa-internal at redhat.com
Depends On: 1705884
Blocks: 1667998, 1668001
Target Milestone: ---
Classification: Community
+++ This bug was initially created as a clone of Bug #1705884 +++
+++ This bug was initially created as a clone of Bug #1668001 +++
Description of problem:
-----------------------
The size of the VM image file as reported from the fuse mount is incorrect.
For the file of size 1 TB, the size of the file on the disk is reported as 8
ZB.
Version-Release number of selected component (if applicable):
-------------------------------------------------------------
upstream master
How reproducible:
------------------
Always
Steps to Reproduce:
-------------------
1. On the Gluster storage domain, create the preallocated disk image of size
1TB
2. Check for the size of the file after its creation has succeesded
Actual results:
---------------
Size of the file is reported as 8 ZB, though the size of the file is 1TB
Expected results:
-----------------
Size of the file should be the same as the size created by the user
Additional info:
----------------
Volume in the question is replica 3 sharded
[root at rhsqa-grafton10 ~]# gluster volume info data
Volume Name: data
Type: Replicate
Volume ID: 7eb49e90-e2b6-4f8f-856e-7108212dbb72
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: rhsqa-grafton10.lab.eng.blr.redhat.com:/gluster_bricks/data/data
Brick2: rhsqa-grafton11.lab.eng.blr.redhat.com:/gluster_bricks/data/data
Brick3: rhsqa-grafton12.lab.eng.blr.redhat.com:/gluster_bricks/data/data
(arbiter)
Options Reconfigured:
performance.client-io-threads: on
nfs.disable: on
transport.address-family: inet
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: off
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
cluster.choose-local: off
client.event-threads: 4
server.event-threads: 4
storage.owner-uid: 36
storage.owner-gid: 36
network.ping-timeout: 30
performance.strict-o-direct: on
cluster.granular-entry-heal: enable
cluster.enable-shared-storage: enable
--- Additional comment from SATHEESARAN on 2019-01-21 16:32:39 UTC ---
Size of the file as reported from the fuse mount:
[root@ ~]# ls -lsah
/rhev/data-center/mnt/glusterSD/rhsqa-grafton10.lab.eng.blr.redhat.com\:_data/bbeee86f-f174-4ec7-9ea3-a0df28709e64/images/0206953c-4850-4969-9dad-15140579d354/eaa5e81d-103c-4ce6-947e-8946806cca1b
8.0Z -rw-rw----. 1 vdsm kvm 1.1T Jan 21 17:14
/rhev/data-center/mnt/glusterSD/rhsqa-grafton10.lab.eng.blr.redhat.com:_data/bbeee86f-f174-4ec7-9ea3-a0df28709e64/images/0206953c-4850-4969-9dad-15140579d354/eaa5e81d-103c-4ce6-947e-8946806cca1b
[root@ ~]# du -shc
/rhev/data-center/mnt/glusterSD/rhsqa-grafton10.lab.eng.blr.redhat.com\:_data/bbeee86f-f174-4ec7-9ea3-a0df28709e64/images/0206953c-4850-4969-9dad-15140579d354/eaa5e81d-103c-4ce6-947e-8946806cca1b
16E
/rhev/data-center/mnt/glusterSD/rhsqa-grafton10.lab.eng.blr.redhat.com:_data/bbeee86f-f174-4ec7-9ea3-a0df28709e64/images/0206953c-4850-4969-9dad-15140579d354/eaa5e81d-103c-4ce6-947e-8946806cca1b
16E total
Note that the disk image is preallocated with 1072GB of space
--- Additional comment from SATHEESARAN on 2019-04-01 19:25:15 UTC ---
(In reply to SATHEESARAN from comment #5)
> (In reply to Krutika Dhananjay from comment #3)
> > Also, do you still have the setup in this state? If so, can I'd like to take
> > a look.
> >
> > -Krutika
>
> Hi Krutika,
>
> The setup is no longer available. Let me recreate the issue and provide you
> the setup
This issue is very easily reproducible. Create a preallocated image on the
replicate volume with sharding enabled.
Use 'qemu-img' to create the VM image.
See the following test:
[root@ ~]# qemu-img create -f raw -o preallocation=falloc /mnt/test/vm1.img 1T
Formatting '/mnt/test/vm1.img', fmt=raw size=1099511627776
preallocation='falloc'
[root@ ]# ls /mnt/test
vm1.img
[root@ ]# ls -lsah vm1.img
8.0Z -rw-r--r--. 1 root root 1.0T Apr 2 00:45 vm1.img
--- Additional comment from Krutika Dhananjay on 2019-04-11 06:07:35 UTC ---
So I tried this locally and I am not hitting the issue -
[root at dhcpxxxxx ~]# qemu-img create -f raw -o preallocation=falloc /mnt/vm1.img
10G
Formatting '/mnt/vm1.img', fmt=raw size=10737418240 preallocation=falloc
[root at dhcpxxxxx ~]# ls -lsah /mnt/vm1.img
10G -rw-r--r--. 1 root root 10G Apr 11 11:26 /mnt/vm1.img
[root at dhcpxxxxx ~]# qemu-img create -f raw -o preallocation=falloc /mnt/vm1.img
30G
Formatting '/mnt/vm1.img', fmt=raw size=32212254720 preallocation=falloc
[root at dhcpxxxxx ~]# ls -lsah /mnt/vm1.img
30G -rw-r--r--. 1 root root 30G Apr 11 11:32 /mnt/vm1.img
Of course, I didn't go beyond 30G due to space constraints on my laptop.
If you could share your setup where you're hitting this bug, I'll take a look.
-Krutika
--- Additional comment from SATHEESARAN on 2019-05-02 05:21:01 UTC ---
(In reply to Krutika Dhananjay from comment #7)
> So I tried this locally and I am not hitting the issue -
>
> [root at dhcpxxxxx ~]# qemu-img create -f raw -o preallocation=falloc
> /mnt/vm1.img 10G
> Formatting '/mnt/vm1.img', fmt=raw size=10737418240 preallocation=falloc
> [root at dhcpxxxxx ~]# ls -lsah /mnt/vm1.img
> 10G -rw-r--r--. 1 root root 10G Apr 11 11:26 /mnt/vm1.img
>
> [root at dhcpxxxxx ~]# qemu-img create -f raw -o preallocation=falloc
> /mnt/vm1.img 30G
> Formatting '/mnt/vm1.img', fmt=raw size=32212254720 preallocation=falloc
> [root at dhcpxxxxx ~]# ls -lsah /mnt/vm1.img
> 30G -rw-r--r--. 1 root root 30G Apr 11 11:32 /mnt/vm1.img
>
> Of course, I didn't go beyond 30G due to space constraints on my laptop.
>
> If you could share your setup where you're hitting this bug, I'll take a
> look.
>
> -Krutika
I could see this very consistenly in two fashions
1. Create VM image >= 1TB
--------------------------
[root at rhsqa-grafton7 test]# qemu-img create -f raw -o preallocation=falloc
vm1.img 10G
Formatting 'vm1.img', fmt=raw size=10737418240 preallocation=falloc
[root@ ]# ls -lsah vm1.img
10G -rw-r--r--. 1 root root 10G May 2 10:30 vm1.img
[root@ ]# qemu-img create -f raw -o preallocation=falloc vm2.img 50G
Formatting 'vm2.img', fmt=raw size=53687091200 preallocation=falloc
[root@ ]# ls -lsah vm2.img
50G -rw-r--r--. 1 root root 50G May 2 10:30 vm2.img
[root@ ]# qemu-img create -f raw -o preallocation=falloc vm3.img 100G
Formatting 'vm3.img', fmt=raw size=107374182400 preallocation=falloc
[root@ ]# ls -lsah vm3.img
100G -rw-r--r--. 1 root root 100G May 2 10:33 vm3.img
[root@ ]# qemu-img create -f raw -o preallocation=falloc vm4.img 500G
Formatting 'vm4.img', fmt=raw size=536870912000 preallocation=falloc
[root@ ]# ls -lsah vm4.img
500G -rw-r--r--. 1 root root 500G May 2 10:33 vm4.img
Once the size reached 1TB, you will see this issue
[root@ ]# qemu-img create -f raw -o preallocation=falloc vm6.img 1T
Formatting 'vm6.img', fmt=raw size=1099511627776 preallocation=falloc
[root@ ]# ls -lsah vm6.img
8.0Z -rw-r--r--. 1 root root 1.0T May 2 10:35 vm6.img <--------
size on disk is too much than expected
2. Recreate the image with the same name
-----------------------------------------
Observe that for the second time, the image is created with the same name
[root@ ]# qemu-img create -f raw -o preallocation=falloc vm1.img 10G
Formatting 'vm1.img', fmt=raw size=10737418240 preallocation=falloc
[root@ ]# ls -lsah vm1.img
10G -rw-r--r--. 1 root root 10G May 2 10:40 vm1.img
[root@ ]# qemu-img create -f raw -o preallocation=falloc vm1.img 20G <--------
The same file name vm1.img is used
Formatting 'vm1.img', fmt=raw size=21474836480 preallocation=falloc
[root@ ]# ls -lsah vm1.img
30G -rw-r--r--. 1 root root 20G May 2 10:40 vm1.img <---------- size on
the disk is 30G, though the file is created with 20G
I will provide setup for the investigation
--- Additional comment from SATHEESARAN on 2019-05-02 05:23:07 UTC ---
The setup details:
-------------------
rhsqa-grafton7.lab.eng.blr.redhat.com ( root/redhat )
volume: data ( replica 3, sharded )
The volume is currently mounted at: /mnt/test
Note: This is the RHVH installation.
@krutika, if you need more info, just ping me in IRC / google chat
--- Additional comment from Krutika Dhananjay on 2019-05-02 10:16:40 UTC ---
Found part of the issue.
It's just a case of integer overflow.
32-bit signed int is being used to store delta between post-stat and pre-stat
block-counts.
The range of numbers for 32-bit signed int is [-2,147,483,648, 2,147,483,647]
whereas the number of blocks allocated
as part of creating a preallocated 1TB file is (1TB/512) = 2,147,483,648 which
is just 1 more than INT_MAX (2,147,483,647)
which spills over to the negative half the scale making it -2,147,483,648.
This number, on being copied to int64 causes the most-significant 32 bits to be
filled with 1 making the block-count equal 554050781183 (or 0xffffffff80000000)
in magnitude.
That's the block-count that gets set on the backend in
trusted.glusterfs.shard.file-size xattr in the block-count segment -
[root at rhsqa-grafton7 data]# getfattr -d -m . -e hex
/gluster_bricks/data/data/vm3.img
getfattr: Removing leading '/' from absolute path names
# file: gluster_bricks/data/data/vm3.img
security.selinux=0x73797374656d5f753a6f626a6563745f723a676c7573746572645f627269636b5f743a733000
trusted.afr.dirty=0x000000000000000000000000
trusted.gfid=0x3faffa7142b74e739f3a82b9359d33e6
trusted.gfid2path.6356251b968111ad=0x30303030303030302d303030302d303030302d303030302d3030303030303030303030312f766d332e696d67
trusted.glusterfs.shard.block-size=0x0000000004000000
trusted.glusterfs.shard.file-size=0x00000100000000000000000000000000ffffffff800000000000000000000000
<-- notice the "ffffffff80000000" in the block-count segment
But ..
[root at rhsqa-grafton7 test]# stat vm3.img
File: ?vm3.img?
Size: 1099511627776 Blocks: 18446744071562067968 IO Block: 131072 regular
file
Device: 29h/41d Inode: 11473626732659815398 Links: 1
Access: (0644/-rw-r--r--) Uid: ( 0/ root) Gid: ( 0/ root)
Context: system_u:object_r:fusefs_t:s0
Access: 2019-05-02 14:11:11.693559069 +0530
Modify: 2019-05-02 14:12:38.245068328 +0530
Change: 2019-05-02 14:15:56.190546751 +0530
Birth: -
stat shows block-count as 18446744071562067968 which is way bigger than
(554050781183 * 512).
In the response path, turns out the block-count further gets assigned to a
uint64 number.
The same number, when expressed as uint64 becomes 18446744071562067968.
18446744071562067968 * 512 is a whopping 8.0 Zettabytes!
This bug wasn't seen earlier because the earlier way of preallocating files
never used fallocate, so the original signed 32 int variable delta_blocks would
never exceed 131072.
Anyway, I'll be soon sending a fix for this.
Sas,
Do you have a single node with at least 1TB free space that you can lend me
where I can test the fix? The bug will only be hit when the image size is >
1TB.
-Krutika
--- Additional comment from Krutika Dhananjay on 2019-05-02 10:18:26 UTC ---
(In reply to Krutika Dhananjay from comment #10)
> Found part of the issue.
Sorry, this not part of the issue but THE issue in its entirety. (That line is
from an older draft I'd composed which I forgot to change after rc'ing the bug)
>
> It's just a case of integer overflow.
> 32-bit signed int is being used to store delta between post-stat and
> pre-stat block-counts.
> The range of numbers for 32-bit signed int is [-2,147,483,648,
> 2,147,483,647] whereas the number of blocks allocated
> as part of creating a preallocated 1TB file is (1TB/512) = 2,147,483,648
> which is just 1 more than INT_MAX (2,147,483,647)
> which spills over to the negative half the scale making it -2,147,483,648.
> This number, on being copied to int64 causes the most-significant 32 bits to
> be filled with 1 making the block-count equal 554050781183 (or
> 0xffffffff80000000) in magnitude.
> That's the block-count that gets set on the backend in
> trusted.glusterfs.shard.file-size xattr in the block-count segment -
>
> [root at rhsqa-grafton7 data]# getfattr -d -m . -e hex
> /gluster_bricks/data/data/vm3.img
> getfattr: Removing leading '/' from absolute path names
> # file: gluster_bricks/data/data/vm3.img
> security.
> selinux=0x73797374656d5f753a6f626a6563745f723a676c7573746572645f627269636b5f7
> 43a733000
> trusted.afr.dirty=0x000000000000000000000000
> trusted.gfid=0x3faffa7142b74e739f3a82b9359d33e6
> trusted.gfid2path.
> 6356251b968111ad=0x30303030303030302d303030302d303030302d303030302d3030303030
> 303030303030312f766d332e696d67
>
> trusted.glusterfs.shard.block-size=0x0000000004000000
> trusted.glusterfs.shard.file-
> size=0x00000100000000000000000000000000ffffffff800000000000000000000000 <--
> notice the "ffffffff80000000" in the block-count segment
>
> But ..
>
> [root at rhsqa-grafton7 test]# stat vm3.img
> File: ?vm3.img?
> Size: 1099511627776 Blocks: 18446744071562067968 IO Block: 131072
> regular file
> Device: 29h/41d Inode: 11473626732659815398 Links: 1
> Access: (0644/-rw-r--r--) Uid: ( 0/ root) Gid: ( 0/ root)
> Context: system_u:object_r:fusefs_t:s0
> Access: 2019-05-02 14:11:11.693559069 +0530
> Modify: 2019-05-02 14:12:38.245068328 +0530
> Change: 2019-05-02 14:15:56.190546751 +0530
> Birth: -
>
> stat shows block-count as 18446744071562067968 which is way bigger than
> (554050781183 * 512).
>
> In the response path, turns out the block-count further gets assigned to a
> uint64 number.
> The same number, when expressed as uint64 becomes 18446744071562067968.
> 18446744071562067968 * 512 is a whopping 8.0 Zettabytes!
>
> This bug wasn't seen earlier because the earlier way of preallocating files
> never used fallocate, so the original signed 32 int variable delta_blocks
> would never exceed 131072.
>
> Anyway, I'll be soon sending a fix for this.
--- Additional comment from Worker Ant on 2019-05-03 06:58:51 UTC ---
REVIEW: https://review.gluster.org/22655 (features/shard: Fix integer overflow
in block count accounting) posted (#1) for review on master by Krutika
Dhananjay
--- Additional comment from Worker Ant on 2019-05-06 10:49:43 UTC ---
REVIEW: https://review.gluster.org/22655 (features/shard: Fix integer overflow
in block count accounting) merged (#2) on master by Xavi Hernandez
--- Additional comment from Worker Ant on 2019-05-08 08:46:18 UTC ---
REVIEW: https://review.gluster.org/22681 (features/shard: Fix block-count
accounting upon truncate to lower size) posted (#1) for review on master by
Krutika Dhananjay
--- Additional comment from Worker Ant on 2019-06-04 07:30:49 UTC ---
REVIEW: https://review.gluster.org/22681 (features/shard: Fix block-count
accounting upon truncate to lower size) merged (#6) on master by Xavi Hernandez
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1667998
[Bug 1667998] Image size as reported from the fuse mount is incorrect
https://bugzilla.redhat.com/show_bug.cgi?id=1668001
[Bug 1668001] Image size as reported from the fuse mount is incorrect
https://bugzilla.redhat.com/show_bug.cgi?id=1705884
[Bug 1705884] Image size as reported from the fuse mount is incorrect
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:29:12 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:29:12 +0000
Subject: [Bugs] [Bug 1705884] Image size as reported from the fuse mount is
incorrect
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1705884
Krutika Dhananjay changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1716871
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1716871
[Bug 1716871] Image size as reported from the fuse mount is incorrect
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:31:39 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:31:39 +0000
Subject: [Bugs] [Bug 1716870] Came up with a script to analyze strace
outputs from bricks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716870
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22816
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:31:40 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:31:40 +0000
Subject: [Bugs] [Bug 1716870] Came up with a script to analyze strace
outputs from bricks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716870
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22816 (extras: Script to analyze strace of
bricks) posted (#1) for review on master by Pranith Kumar Karampuri
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:32:31 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:32:31 +0000
Subject: [Bugs] [Bug 1622665] clang-scan report: glusterfs issues
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1622665
--- Comment #92 from Worker Ant ---
REVIEW: https://review.gluster.org/22775 (across: clang-scan: fix NULL
dereferencing warnings) merged (#3) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:35:38 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:35:38 +0000
Subject: [Bugs] [Bug 1716871] Image size as reported from the fuse mount is
incorrect
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716871
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22817
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:35:39 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:35:39 +0000
Subject: [Bugs] [Bug 1716871] Image size as reported from the fuse mount is
incorrect
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716871
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22817 (features/shard: Fix integer overflow
in block count accounting) posted (#1) for review on release-6 by Krutika
Dhananjay
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:37:46 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:37:46 +0000
Subject: [Bugs] [Bug 1716875] New: Inode Unref Assertion failed: inode->ref
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716875
Bug ID: 1716875
Summary: Inode Unref Assertion failed: inode->ref
Product: GlusterFS
Version: 4.1
Hardware: x86_64
OS: Linux
Status: NEW
Component: gluster-smb
Severity: urgent
Assignee: bugs at gluster.org
Reporter: ryan at magenta.tv
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Created attachment 1577014
--> https://bugzilla.redhat.com/attachment.cgi?id=1577014&action=edit
Client log from Gluster VFS client showing high RAM usage
Description of problem:
Samba using huge amounts of memory (5GB) per client thread.
Upon checking the Gluster client logs, it's filled with messages such as:
[2019-04-24 07:44:33.607834] E [inode.c:484:__inode_unref]
(-->/lib64/libglusterfs.so.0(gf_dirent_entry_free+0x2b) [0x7ff0a24d555b]
-->/lib64/libglusterfs.so.0(inode_unref+0x21) [0x7ff0a24b9921]
-->/lib64/libglusterfs.so.0(+0x35156) [0x7ff0a24b9156] ) 0-: Assertion failed:
inode->ref
[2019-04-30 13:16:47.169047] E [timer.c:37:gf_timer_call_after]
(-->/lib64/libglusterfs.so.0(+0x33bec) [0x7ff09d875bec]
-->/lib64/libgfrpc.so.0(+0xde88) [0x7ff09dd7ae88]
-->/lib64/libglusterfs.so.0(gf_timer_call_after+0x229) [0x7ff09d875fa9] )
0-timer: Either ctx is NULL or ctx cleanup started [Invalid argument]
[2019-05-28 17:47:28.655550] E [MSGID: 140003] [nl-cache.c:777:nlc_init]
0-mcv01-nl-cache: Initing the global timer wheel failed
[2019-05-28 17:47:28.655873] E [MSGID: 101019] [xlator.c:720:xlator_init]
0-mcv01-nl-cache: Initialization of volume 'mcv01-nl-cache' failed, review your
volfile again
[2019-05-28 17:47:28.655887] E [MSGID: 101066]
[graph.c:367:glusterfs_graph_init] 0-mcv01-nl-cache: initializing translator
failed
[2019-05-28 17:47:28.655894] E [MSGID: 101176]
[graph.c:738:glusterfs_graph_activate] 0-graph: init failed
[2019-05-28 17:47:28.655972] E [MSGID: 104007]
[glfs-mgmt.c:744:glfs_mgmt_getspec_cbk] 0-glfs-mgmt: failed to fetch volume
file (key:mcv01) [Invalid argument]
Version-Release number of selected component (if applicable):
Gluster 4.1.7
How reproducible:
Unknown
Steps to Reproduce:
Unsure how to reproduce, only seen this in one environment currently
Actual results:
All system memory and swap is exhausted. SMBD processes do not get killed off
when main SMB service is stopped, where as usually they do.
Expected results:
System resources are freed up and errors are not present in logs.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 10:54:31 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 10:54:31 +0000
Subject: [Bugs] [Bug 1705351] glusterfsd crash after days of running
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1705351
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags|needinfo?(jahernan at redhat.c |needinfo?(waza123 at inbox.lv)
|om) |
--- Comment #5 from Xavi Hernandez ---
Sorry for the late answer. I've checked the core dump and it seems to belong to
a glusterfs 3.10.10. This is a very old version and it's already EOL. Is it
possible to upgrade to a newer supported version and check if it works ?
At first sight I don't see a similar bug, but many things have changed since
then.
If you are unable to upgrade, let me know which version of operating system are
you using and which source you use to install gluster packages so that I can
find the appropriate symbols to analyze the core.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 11:08:19 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 11:08:19 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22818
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 11:08:20 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 11:08:20 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
--- Comment #55 from Worker Ant ---
REVIEW: https://review.gluster.org/22818 (tests/geo-rep: Add geo-rep cli
testcases) posted (#1) for review on master by Kotresh HR
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 11:32:39 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 11:32:39 +0000
Subject: [Bugs] [Bug 1716871] Image size as reported from the fuse mount is
incorrect
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716871
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22819
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 11:32:40 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 11:32:40 +0000
Subject: [Bugs] [Bug 1716871] Image size as reported from the fuse mount is
incorrect
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716871
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22819 (features/shard: Fix block-count
accounting upon truncate to lower size) posted (#1) for review on release-6 by
Krutika Dhananjay
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 11:35:04 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 11:35:04 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22820
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 11:35:05 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 11:35:05 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #679 from Worker Ant ---
REVIEW: https://review.gluster.org/22820 ([WIP][RFC]dict: use fixed 'hash' for
keys that are fixed strings.) posted (#1) for review on master by Yaniv Kaul
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 12:27:48 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 12:27:48 +0000
Subject: [Bugs] [Bug 1468510] Keep all Debug level log in circular in-memory
buffer
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1468510
Yaniv Kaul changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |aspandey at redhat.com
Flags| |needinfo?(aspandey at redhat.c
| |om)
--- Comment #13 from Yaniv Kaul ---
Since it was not implemented for ~2 years, shall we close it?
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 13:38:03 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 13:38:03 +0000
Subject: [Bugs] [Bug 1716979] New: Multiple disconnect events being
propagated for the same child
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716979
Bug ID: 1716979
Summary: Multiple disconnect events being propagated for the
same child
Product: GlusterFS
Version: mainline
OS: Linux
Status: NEW
Component: rpc
Keywords: Regression
Severity: high
Priority: high
Assignee: bugs at gluster.org
Reporter: rgowdapp at redhat.com
CC: amukherj at redhat.com, bugs at gluster.org,
ravishankar at redhat.com, rgowdapp at redhat.com,
rhinduja at redhat.com, rhs-bugs at redhat.com,
sankarshan at redhat.com, sheggodu at redhat.com
Depends On: 1703423
Target Milestone: ---
Classification: Community
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1703423
[Bug 1703423] Multiple disconnect events being propagated for the same child
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 13:39:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 13:39:06 +0000
Subject: [Bugs] [Bug 1716979] Multiple disconnect events being propagated
for the same child
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716979
--- Comment #1 from Raghavendra G ---
Issue was reported upstream by a user via
https://github.com/gluster/glusterfs/issues/648
I'm seeing that if I kill a brick in a replica 3 system, AFR keeps getting
child_down event repeatedly for the same child.
Version-Release number of selected component (if applicable):
master (source install)
How reproducible:
Always.
Steps to Reproduce:
1. Create a replica 3 volume and start it.
2. Put a break point in __afr_handle_child_down_event() in glustershd process.
3. Kill any one brick.
Actual results:
The break point keeps getting hit once every 3 seconds or so repeatedly.
Expected results:
Only 1 event per one disconnect.
Additional info:
I haven't checked if the same happens for GF_EVENT_CHILD_UP as well. I think
this is regression that needs to be fixed. If this is not a bug please feel
free to close stating why.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 13:52:16 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 13:52:16 +0000
Subject: [Bugs] [Bug 1716979] Multiple disconnect events being propagated
for the same child
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716979
Raghavendra G changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
--- Comment #2 from Raghavendra G ---
The multiple disconnect events are due to reconnect/disconnect to glusterd
(port 24007). rpc/clnt has a reconnect feature which tries to reconnect to a
disconnected brick and client connection to brick is a two step process:
1. connect to glusterd, get brick port then disconnect
2. connect to brick
In this case step 1 would be successful and step 2 won't happen as glusterd
wouldn't send back the brick port (as brick is dead). Nevertheless there is a
chain of connect/disconnect (to glusterd) at rpc layer and these are valid
steps as we need reconnection logic. However subsequent disconnect events were
prevented from reaching parents of protocol/client as it remembered which was
the last sent and if current event is the same as last event, it would skip
notification. Before Halo replication feature -
https://review.gluster.org/16177, last_sent_event for this test case would be
GF_EVENT_DISCONNECT and hence subsequent disconnects were skipped notification
to parent xlators. But Halo replication introduced another event
GF_EVENT_CHILD_PING which gets notified to parents of protocol/client whenever
there is a successful ping response. In this case, the successful ping response
would be from glusterd and would change conf->last_sent_event to
GF_EVENT_CHILD_PING. This made subsequent disconnect events are not skipped.
A patch to propagate GF_EVENT_CHILD_PING only after a successful handshake
prevents spurious CHILD_DOWN events to afr. However, I am not sure whether this
breaks Halo replication. Would request afr team members comment on the patch
(I'll post shortly).
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 13:53:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 13:53:06 +0000
Subject: [Bugs] [Bug 1468510] Keep all Debug level log in circular in-memory
buffer
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1468510
Ashish Pandey changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |CLOSED
Resolution|--- |WONTFIX
Flags|needinfo?(aspandey at redhat.c |
|om) |
Last Closed| |2019-06-04 13:53:06
--- Comment #14 from Ashish Pandey ---
Yes, We can closing it now.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 14:10:12 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 14:10:12 +0000
Subject: [Bugs] [Bug 1716979] Multiple disconnect events being propagated
for the same child
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716979
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22821
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Tue Jun 4 14:10:13 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Tue, 04 Jun 2019 14:10:13 +0000
Subject: [Bugs] [Bug 1716979] Multiple disconnect events being propagated
for the same child
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716979
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--- Comment #3 from Worker Ant ---
REVIEW: https://review.gluster.org/22821 (protocol/client: propagte
GF_EVENT_CHILD_PING only after a successful handshake) posted (#1) for review
on master by Raghavendra G
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 05:52:32 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 05:52:32 +0000
Subject: [Bugs] [Bug 1717282] New: ec ignores lock contention notifications
for partially acquired locks
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717282
Bug ID: 1717282
Summary: ec ignores lock contention notifications for partially
acquired locks
Product: GlusterFS
Version: 5
Status: NEW
Component: disperse
Assignee: bugs at gluster.org
Reporter: jahernan at redhat.com
CC: bugs at gluster.org
Depends On: 1708156
Blocks: 1714172
Target Milestone: ---
Classification: Community
+++ This bug was initially created as a clone of Bug #1708156 +++
Description of problem:
When an inodelk is being acquired, it could happen that some bricks have
already granted the lock while others don't. From the point of view of ec, the
lock is not yet acquired.
If at this point one of the bricks that has already granted the lock receives
another inodelk request, it will send a contention notification to ec.
Currently ec ignores those notifications until the lock is fully acquired. This
means than once ec acquires the lock on all bricks, it won't be released
immediately when eager-lock is used.
Version-Release number of selected component (if applicable): mainline
How reproducible:
Very frequently when there are multiple concurrent operations on same directory
Steps to Reproduce:
1. Create a disperse volume
2. Mount it from several clients
3. Create few files on a directory
4. Do 'ls' of that directory at the same time from all clients
Actual results:
Some 'ls' take several seconds to complete
Expected results:
All 'ls' should complete in less than a second
Additional info:
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1708156
[Bug 1708156] ec ignores lock contention notifications for partially acquired
locks
https://bugzilla.redhat.com/show_bug.cgi?id=1714172
[Bug 1714172] ec ignores lock contention notifications for partially acquired
locks
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 05:52:32 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 05:52:32 +0000
Subject: [Bugs] [Bug 1708156] ec ignores lock contention notifications for
partially acquired locks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1708156
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1717282
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1717282
[Bug 1717282] ec ignores lock contention notifications for partially acquired
locks
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Wed Jun 5 05:52:32 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 05:52:32 +0000
Subject: [Bugs] [Bug 1714172] ec ignores lock contention notifications for
partially acquired locks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714172
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
Depends On| |1717282
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1717282
[Bug 1717282] ec ignores lock contention notifications for partially acquired
locks
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Wed Jun 5 05:54:54 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 05:54:54 +0000
Subject: [Bugs] [Bug 1717282] ec ignores lock contention notifications for
partially acquired locks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717282
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22822
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 05:54:55 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 05:54:55 +0000
Subject: [Bugs] [Bug 1717282] ec ignores lock contention notifications for
partially acquired locks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717282
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22822 (cluster/ec: honor contention
notifications for partially acquired locks) posted (#1) for review on release-5
by Xavi Hernandez
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 06:04:25 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 06:04:25 +0000
Subject: [Bugs] [Bug 1697986] GlusterFS 5.7 tracker
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1697986
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |jahernan at redhat.com
Depends On| |1717282
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1717282
[Bug 1717282] ec ignores lock contention notifications for partially acquired
locks
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 06:04:25 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 06:04:25 +0000
Subject: [Bugs] [Bug 1717282] ec ignores lock contention notifications for
partially acquired locks
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717282
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1697986 (glusterfs-5.7)
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1697986
[Bug 1697986] GlusterFS 5.7 tracker
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 07:21:00 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 07:21:00 +0000
Subject: [Bugs] [Bug 1716830] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-05 07:21:00
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22813 (cluster/dht: Fix directory perms
during selfheal) merged (#2) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Wed Jun 5 07:21:01 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 07:21:01 +0000
Subject: [Bugs] [Bug 1716848] DHT: directory permissions are wiped out
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716848
Bug 1716848 depends on bug 1716830, which changed state.
Bug 1716830 Summary: DHT: directory permissions are wiped out
https://bugzilla.redhat.com/show_bug.cgi?id=1716830
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Wed Jun 5 16:34:01 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 16:34:01 +0000
Subject: [Bugs] [Bug 1693693] GlusterFS 4.1.9 tracker
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693693
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22826
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 16:34:02 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 16:34:02 +0000
Subject: [Bugs] [Bug 1693693] GlusterFS 4.1.9 tracker
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693693
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22826 (doc: Added release notes for 4.1.9)
posted (#1) for review on release-4.1 by hari gowtham
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 17:19:59 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 17:19:59 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22827
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Wed Jun 5 17:20:00 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Wed, 05 Jun 2019 17:20:00 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #680 from Worker Ant ---
REVIEW: https://review.gluster.org/22827 ([WIP]cli: defer create_frame() (and
dict creation) to later stages.) posted (#1) for review on master by Yaniv Kaul
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 04:05:33 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 04:05:33 +0000
Subject: [Bugs] [Bug 1655201] dictionary leak at the time of destroying graph
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1655201
Mohit Agrawal changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |MODIFIED
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 04:06:39 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 04:06:39 +0000
Subject: [Bugs] [Bug 1711827] test case bug-1399598-uss-with-ssl.t is
generating crash
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1711827
Mohit Agrawal changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |MODIFIED
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 05:18:08 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 05:18:08 +0000
Subject: [Bugs] [Bug 1693692] Increase code coverage from regression tests
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1693692
--- Comment #56 from Worker Ant ---
REVIEW: https://review.gluster.org/22818 (tests/geo-rep: Add geo-rep cli
testcases) merged (#5) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 06:28:32 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:28:32 +0000
Subject: [Bugs] [Bug 1717754] New: Enabled features.locks-notify-contention
by default
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717754
Bug ID: 1717754
Summary: Enabled features.locks-notify-contention by default
Product: GlusterFS
Version: mainline
Status: NEW
Component: locks
Assignee: bugs at gluster.org
Reporter: jahernan at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
Currently 'features.locks-notify-contention' is disabled by default. This
option, when enabled, instructs the locks xlator to send an upcall notification
to the current owner of a lock whenever another client tries to acquire a
conflicting lock.
Both AFR and EC support this notification and react by releasing the lock as
soon as possible. This is extremely useful when eager-lock is enabled (it is by
default) because it allows AFR and EC to use it to improve performance but we
don't loose performance on other clients when access to the same resource is
required.
Since eager-lock is enabled by default, it doesn't make sense to keep
contention notification disabled.
Version-Release number of selected component (if applicable): mainline
How reproducible:
Steps to Reproduce:
1.
2.
3.
Actual results:
Expected results:
Additional info:
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 06:29:01 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:29:01 +0000
Subject: [Bugs] [Bug 1717754] Enabled features.locks-notify-contention by
default
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717754
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
Assignee|bugs at gluster.org |jahernan at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 06:31:14 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:31:14 +0000
Subject: [Bugs] [Bug 1717754] Enabled features.locks-notify-contention by
default
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717754
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22828
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 06:31:15 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:31:15 +0000
Subject: [Bugs] [Bug 1717754] Enabled features.locks-notify-contention by
default
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717754
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22828 (locks: enable notify-contention by
default) posted (#1) for review on master by Xavi Hernandez
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 06:35:37 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:35:37 +0000
Subject: [Bugs] [Bug 1717754] Enable features.locks-notify-contention by
default
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717754
Xavi Hernandez changed:
What |Removed |Added
----------------------------------------------------------------------------
Summary|Enabled |Enable
|features.locks-notify-conte |features.locks-notify-conte
|ntion by default |ntion by default
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 06:41:14 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:41:14 +0000
Subject: [Bugs] [Bug 1717757] New: BItrot: Segmentation Fault if bitrot stub
do signature
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
Bug ID: 1717757
Summary: BItrot: Segmentation Fault if bitrot stub do signature
Product: GlusterFS
Version: 5
Status: NEW
Component: bitrot
Assignee: bugs at gluster.org
Reporter: david.spisla at iternity.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Docs Contact: bugs at gluster.org
Created attachment 1577785
--> https://bugzilla.redhat.com/attachment.cgi?id=1577785&action=edit
backtrace of /usr/sbin/glusterfsd
Description of problem:
Setup: 2-Node VM Cluster with a Replica 2 Volume
After doing several "wild" write and delete operations from a Win Client, one
of the brick crashes. The crash report says:
[2019-06-05 09:05:05.137156] I [MSGID: 139001]
[posix-acl.c:263:posix_acl_log_permit_denied] 0-archive1-access-control:
client: CTX_ID:fcab5e67-b9d9-4b72-8c15-f29de2084af3-GRAPH_ID:0-PID:189
16-HOST:fs-detlefh-c1-n2-PC_NAME:archive1-client-0-RECON_NO:-0, gfid:
494b42ad-7e40-4e27-8878-99387a80b5dc, req(uid:2000,gid:2000,perm:3,ngrps:1),
ctx(uid:0,gid:0,in-groups:0,perm:755,update
d-fop:LOOKUP, acl:-) [Permission denied]
pending frames:
frame : type(0) op(0)
frame : type(0) op(23)
patchset: git://git.gluster.org/glusterfs.git
signal received: 11
time of crash:
2019-06-05 09:05:05
configuration details:
argp 1
backtrace 1
dlfcn 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 5.5
/usr/lib64/libglusterfs.so.0(+0x2764c)[0x7f89faa7264c]
/usr/lib64/libglusterfs.so.0(gf_print_trace+0x306)[0x7f89faa7cd26]
/lib64/libc.so.6(+0x361a0)[0x7f89f9c391a0]
/usr/lib64/glusterfs/5.5/xlator/features/bitrot-stub.so(+0x13441)[0x7f89f22ae441]
/usr/lib64/libglusterfs.so.0(default_fsetxattr+0xce)[0x7f89faaf9f8e]
/usr/lib64/glusterfs/5.5/xlator/features/locks.so(+0x22636)[0x7f89f1e68636]
/usr/lib64/libglusterfs.so.0(default_fsetxattr+0xce)[0x7f89faaf9f8e]
/usr/lib64/libglusterfs.so.0(syncop_fsetxattr+0x26b)[0x7f89faab319b]
/usr/lib64/glusterfs/5.5/xlator/features/worm.so(+0xa901)[0x7f89f1c3d901]
/usr/lib64/glusterfs/5.5/xlator/features/locks.so(+0x11b66)[0x7f89f1e57b66]
/usr/lib64/glusterfs/5.5/xlator/features/access-control.so(+0xaebe)[0x7f89f208febe]
/usr/lib64/glusterfs/5.5/xlator/features/locks.so(+0xb081)[0x7f89f1e51081]
/usr/lib64/glusterfs/5.5/xlator/features/worm.so(+0x8c23)[0x7f89f1c3bc23]
/usr/lib64/glusterfs/5.5/xlator/features/read-only.so(+0x4e30)[0x7f89f1a2de30]
/usr/lib64/glusterfs/5.5/xlator/features/leases.so(+0xa444)[0x7f89f181b444]
/usr/lib64/glusterfs/5.5/xlator/features/upcall.so(+0x10a68)[0x7f89f1600a68]
/usr/lib64/libglusterfs.so.0(default_create_resume+0x212)[0x7f89fab10132]
/usr/lib64/libglusterfs.so.0(call_resume_wind+0x2cf)[0x7f89faa97e5f]
/usr/lib64/libglusterfs.so.0(call_resume+0x75)[0x7f89faa983a5]
/usr/lib64/glusterfs/5.5/xlator/performance/io-threads.so(+0x6088)[0x7f89f13e7088]
/lib64/libpthread.so.0(+0x7569)[0x7f89f9fc4569]
/lib64/libc.so.6(clone+0x3f)[0x7f89f9cfb9af]
---------
Version-Release number of selected component (if applicable): v5.5
Additional info:
The backtrace shows that there is a Nulllpointer for *fd in br_stub_fsetxattr:
Thread 1 (Thread 0x7f89f0099700 (LWP 2171)):
#0 br_stub_fsetxattr (frame=0x7f89b846a6e8, this=0x7f89ec015c00, fd=0x0,
dict=0x7f89b84e9ad8, flags=0, xdata=0x0) at bit-rot-stub.c:1328
ret = 0
val = 0
sign = 0x0
priv = 0x7f89ec07ed60
op_errno = 22
__FUNCTION__ = "br_stub_fsetxattr"
This results in a segmentation fault in line 1328 of bit-rot_stub.c :
if (!IA_ISREG(fd->inode->ia_type))
goto wind;
The bitrot-stub wants to signate a file but the corresponding fd is a
Nullpointer.
The full backtrace is attached!!!
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Thu Jun 6 06:44:04 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:44:04 +0000
Subject: [Bugs] [Bug 1706842] Hard Failover with Samba and Glusterfs fails
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1706842
--- Comment #4 from david.spisla at iternity.com ---
Additional Information: My setup was a 4-Node Cluster with VM machines (VmWare)
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 06:57:25 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:57:25 +0000
Subject: [Bugs] [Bug 1717757] BItrot: Segmentation Fault if bitrot stub do
signature
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
Amar Tumballi changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
CC| |atumball at redhat.com
--- Comment #1 from Amar Tumballi ---
Not sure why this happened, because, for bitrot, a fsetxattr() call shouldn't
come at all if fd is NULL. It should have been prevented at higher level
itself.
I found the reason after digging a bit. Ideally, in case of failure (here,
worm_create_cbk() received -1, which means fd is NULL), one shouldn't consume
fd and call fsetxattr(). If there is a need to do a xattr op in failure, then
one should call setxattr with 'loc' passed in create() call. (you can store it
in local).
----
#0 br_stub_fsetxattr (frame=0x7f89b846a6e8, this=0x7f89ec015c00, fd=0x0,
dict=0x7f89b84e9ad8, flags=0, xdata=0x0) at bit-rot-stub.c:1328
ret = 0
val = 0
sign = 0x0
priv = 0x7f89ec07ed60
op_errno = 22
__FUNCTION__ = "br_stub_fsetxattr"
#1 0x00007f89faaf9f8e in default_fsetxattr () from
/usr/lib64/libglusterfs.so.0
No symbol table info available.
#2 0x00007f89f1e68636 in pl_fsetxattr (frame=0x7f89b825ab48,
this=0x7f89ec0194a0, fd=0x0, dict=0x7f89b84e9ad8, flags=0, xdata=0x0) at
posix.c:1566
_new = 0x7f89b846a6e8
old_THIS = 0x7f89ec0194a0
next_xl_fn = 0x7f89faaf9ec0
tmp_cbk = 0x7f89f1e56680
op_ret =
op_errno = 0
lockinfo_buf = 0x0
len = 0
__FUNCTION__ = "pl_fsetxattr"
#3 0x00007f89faaf9f8e in default_fsetxattr () from
/usr/lib64/libglusterfs.so.0
No symbol table info available.
#4 0x00007f89faab319b in syncop_fsetxattr () from /usr/lib64/libglusterfs.so.0
No symbol table info available.
#5 0x00007f89f1c3d901 in worm_create_cbk (frame=frame at entry=0x7f89b8302fe8,
cookie=, this=, op_ret=op_ret at entry=-1,
op_errno=op_errno at entry=13,
fd=fd at entry=0x0, inode=0x0, buf=0x0, preparent=0x0, postparent=0x0,
xdata=0x0) at worm.c:492
ret = 0
priv = 0x7f89ec074b38
dict = 0x7f89b84e9ad8
__FUNCTION__ = "worm_create_cbk"
----
Hopefully this helps.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Thu Jun 6 06:59:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:59:29 +0000
Subject: [Bugs] [Bug 1717757] BItrot: Segmentation Fault if bitrot stub do
signature
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
--- Comment #2 from Amar Tumballi ---
Can you check if below works?
diff --git a/xlators/features/read-only/src/worm.c
b/xlators/features/read-only/src/worm.c
index cc3d15b8b2..6b44eae966 100644
--- a/xlators/features/read-only/src/worm.c
+++ b/xlators/features/read-only/src/worm.c
@@ -431,7 +431,7 @@ worm_create_cbk(call_frame_t *frame, void *cookie, xlator_t
*this,
priv = this->private;
GF_ASSERT(priv);
- if (priv->worm_file) {
+ if (priv->worm_file && (op_ret >= 0)) {
dict = dict_new();
if (!dict) {
gf_log(this->name, GF_LOG_ERROR,
----
Great if you can confirm this.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Thu Jun 6 06:59:51 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 06:59:51 +0000
Subject: [Bugs] [Bug 1717757] BItrot: Segmentation Fault if bitrot stub do
signature
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
Amar Tumballi changed:
What |Removed |Added
----------------------------------------------------------------------------
Priority|unspecified |medium
Assignee|bugs at gluster.org |atumball at redhat.com
Severity|unspecified |high
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Thu Jun 6 07:08:49 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 07:08:49 +0000
Subject: [Bugs] [Bug 1717757] BItrot: Segmentation Fault if bitrot stub do
signature
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
--- Comment #3 from david.spisla at iternity.com ---
I will check it!
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Thu Jun 6 07:12:54 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 07:12:54 +0000
Subject: [Bugs] [Bug 1714536] geo-rep: With heavy rename workload geo-rep
log if flooded
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714536
Vivek Das changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |vdas at redhat.com
Blocks| |1696809
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 07:12:59 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 07:12:59 +0000
Subject: [Bugs] [Bug 1714536] geo-rep: With heavy rename workload geo-rep
log if flooded
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714536
RHEL Product and Program Management changed:
What |Removed |Added
----------------------------------------------------------------------------
Rule Engine Rule| |Gluster: Auto pm_ack for
| |dev&qe approved in-flight
| |RHGS3.5 BZs
Rule Engine Rule| |665
Target Release|--- |RHGS 3.5.0
Rule Engine Rule| |666
Rule Engine Rule| |327
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 07:35:40 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 07:35:40 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22829
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 07:56:57 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 07:56:57 +0000
Subject: [Bugs] [Bug 1717782] gluster v get all still showing
storage.fips-mode-rchecksum off
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717782
Ravishankar N changed:
What |Removed |Added
----------------------------------------------------------------------------
Keywords| |Triaged
Status|NEW |ASSIGNED
Assignee|bugs at gluster.org |ravishankar at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 08:01:17 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 08:01:17 +0000
Subject: [Bugs] [Bug 1717782] gluster v get all still showing
storage.fips-mode-rchecksum off
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717782
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22830
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 08:01:18 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 08:01:18 +0000
Subject: [Bugs] [Bug 1717782] gluster v get all still showing
storage.fips-mode-rchecksum off
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717782
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22830 (glusterd: store fips-mode-rchecksum
option in the info file) posted (#1) for review on master by Ravishankar N
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 07:35:41 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 07:35:41 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #681 from Worker Ant ---
REVIEW: https://review.gluster.org/22829 (tests/utils: Fix py2/py3
changelogparser.py compatibility) posted (#1) for review on master by Kotresh
HR
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 09:21:37 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:21:37 +0000
Subject: [Bugs] [Bug 1717819] New: Changes to self-heal logic w.r.t.
detecting metadata split-brains
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717819
Bug ID: 1717819
Summary: Changes to self-heal logic w.r.t. detecting metadata
split-brains
Product: GlusterFS
Version: mainline
Status: ASSIGNED
Component: replicate
Assignee: ksubrahm at redhat.com
Reporter: ksubrahm at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
We currently don't have a roll-back/undoing of post-ops if quorum is not met.
Though the FOP is still unwound with failure, the xattrs remain on the disk.
Due to these partial post-ops and partial heals (healing only when 2 bricks are
up), we can end up in metadata split-brain purely from the afr xattrs point of
view i.e each brick is blamed by atleast one of the others for metadata. These
scenarios are hit when there is frequent connect/disconnect of the client/shd
to the bricks.
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 09:34:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:34:29 +0000
Subject: [Bugs] [Bug 1717819] Changes to self-heal logic w.r.t. detecting
metadata split-brains
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717819
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22831
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 09:34:30 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:34:30 +0000
Subject: [Bugs] [Bug 1717819] Changes to self-heal logic w.r.t. detecting
metadata split-brains
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717819
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22831 (Cluster/afr: Don't treat all bricks
having metadata pending as split-brain) posted (#1) for review on master by
Karthik U S
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 09:35:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:35:29 +0000
Subject: [Bugs] [Bug 1717824] New: Fencing: Added the tcmu-runner ALUA
feature support but after one of node is rebooted the glfs_file_lock() get
stucked
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717824
Bug ID: 1717824
Summary: Fencing: Added the tcmu-runner ALUA feature support
but after one of node is rebooted the glfs_file_lock()
get stucked
Product: GlusterFS
Version: mainline
Status: NEW
Component: locks
Assignee: bugs at gluster.org
Reporter: xiubli at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
In Glusterfs, we have support the fencing feature support. With this we can
suppor the ALUA feature in LIO/TCMU now.
The fencing doc:
https://review.gluster.org/#/c/glusterfs-specs/+/21925/6/accepted/fencing.md
The fencing test example:
https://review.gluster.org/#/c/glusterfs/+/21496/12/tests/basic/fencing/fence-basic.c
The LIO/tcmu-runner PR of supporting the ALUA is :
https://github.com/open-iscsi/tcmu-runner/pull/554.
But currently when testing it based the above PR in tcmu-runner by shutting
down of the HA node, and start it after 2~3 minutes, in all the HA nodes we can
see that the glfs_file_lock() get stucked, the following is from the
/var/log/tcmu-runner.log:
====
2019-06-06 13:50:15.755 1316 [DEBUG] tcmu_acquire_dev_lock:388 glfs/block3:
lock call state 2 retries 0. tag 65535 reopen 0
2019-06-06 13:50:15.757 1316 [DEBUG] tcmu_acquire_dev_lock:440 glfs/block3:
lock call done. lock state 1
2019-06-06 13:50:55.845 1316 [DEBUG] tcmu_acquire_dev_lock:388 glfs/block4:
lock call state 2 retries 0. tag 65535 reopen 0
2019-06-06 13:50:55.847 1316 [DEBUG] tcmu_acquire_dev_lock:440 glfs/block4:
lock call done. lock state 1
2019-06-06 13:57:50.102 1315 [DEBUG] tcmu_acquire_dev_lock:388 glfs/block3:
lock call state 2 retries 0. tag 65535 reopen 0
2019-06-06 13:57:50.103 1315 [DEBUG] tcmu_acquire_dev_lock:440 glfs/block3:
lock call done. lock state 1
2019-06-06 13:57:50.121 1315 [DEBUG] tcmu_acquire_dev_lock:388 glfs/block4:
lock call state 2 retries 0. tag 65535 reopen 0
2019-06-06 13:57:50.132 1315 [DEBUG] tcmu_acquire_dev_lock:440 glfs/block4:
lock call done. lock state 1
2019-06-06 14:09:03.654 1328 [DEBUG] tcmu_acquire_dev_lock:388 glfs/block3:
lock call state 2 retries 0. tag 65535 reopen 0
2019-06-06 14:09:03.662 1328 [DEBUG] tcmu_acquire_dev_lock:440 glfs/block3:
lock call done. lock state 1
2019-06-06 14:09:06.700 1328 [DEBUG] tcmu_acquire_dev_lock:388 glfs/block4:
lock call state 2 retries 0. tag 65535 reopen 0
====
The lock operation is never returned.
I am using the following glusterfs built by myself:
# rpm -qa|grep glusterfs
glusterfs-extra-xlators-7dev-0.0.el7.x86_64
glusterfs-api-devel-7dev-0.0.el7.x86_64
glusterfs-7dev-0.0.el7.x86_64
glusterfs-server-7dev-0.0.el7.x86_64
glusterfs-cloudsync-plugins-7dev-0.0.el7.x86_64
glusterfs-resource-agents-7dev-0.0.el7.noarch
glusterfs-api-7dev-0.0.el7.x86_64
glusterfs-devel-7dev-0.0.el7.x86_64
glusterfs-regression-tests-7dev-0.0.el7.x86_64
glusterfs-gnfs-7dev-0.0.el7.x86_64
glusterfs-client-xlators-7dev-0.0.el7.x86_64
glusterfs-geo-replication-7dev-0.0.el7.x86_64
glusterfs-debuginfo-7dev-0.0.el7.x86_64
glusterfs-fuse-7dev-0.0.el7.x86_64
glusterfs-events-7dev-0.0.el7.x86_64
glusterfs-libs-7dev-0.0.el7.x86_64
glusterfs-cli-7dev-0.0.el7.x86_64
glusterfs-rdma-7dev-0.0.el7.x86_64
How reproducible:
30%.
Steps to Reproduce:
1. create one rep volume(HA >= 2) with the mandantary lock enabled
2. create one gluster-blockd target
3. login and do the fio in the client node
4. shutdown one of the HA nodes, and wait 2 ~3 minutes and start it again
Actual results:
all the time the fio couldn't recovery and the rw BW will be 0kb/s, and we can
see tons of log from /var/log/tcmu-runnner.log file:
2019-06-06 15:01:06.641 1328 [DEBUG] alua_implicit_transition:561 glfs/block4:
Lock acquisition operation is already in process.
2019-06-06 15:01:06.648 1328 [DEBUG_SCSI_CMD] tcmu_cdb_print_info:353
glfs/block4: 28 0 0 3 1f 80 0 0 8 0
2019-06-06 15:01:06.648 1328 [DEBUG] alua_implicit_transition:561 glfs/block4:
Lock acquisition operation is already in process.
2019-06-06 15:01:06.655 1328 [DEBUG_SCSI_CMD] tcmu_cdb_print_info:353
glfs/block4: 28 0 0 3 1f 80 0 0 8 0
2019-06-06 15:01:06.655 1328 [DEBUG] alua_implicit_transition:561 glfs/block4:
Lock acquisition operation is already in process.
2019-06-06 15:01:06.661 1328 [DEBUG_SCSI_CMD] tcmu_cdb_print_info:353
glfs/block4: 28 0 0 3 1f 80 0 0 8 0
2019-06-06 15:01:06.662 1328 [DEBUG] alua_implicit_transition:561 glfs/block4:
Lock acquisition operation is already in process.
Expected results:
just before the shutdown node is up, the fio could be recovery.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 09:36:21 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:36:21 +0000
Subject: [Bugs] [Bug 1717824] Fencing: Added the tcmu-runner ALUA feature
support but after one of node is rebooted the glfs_file_lock() get stucked
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717824
Susant Kumar Palai changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |spalai at redhat.com
Assignee|bugs at gluster.org |spalai at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 09:39:50 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:39:50 +0000
Subject: [Bugs] [Bug 1717824] Fencing: Added the tcmu-runner ALUA feature
support but after one of node is rebooted the glfs_file_lock() get stucked
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717824
--- Comment #1 from Xiubo Li ---
Created attachment 1577819
--> https://bugzilla.redhat.com/attachment.cgi?id=1577819&action=edit
pstack of on node rhel3
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 09:40:16 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:40:16 +0000
Subject: [Bugs] [Bug 1717824] Fencing: Added the tcmu-runner ALUA feature
support but after one of node is rebooted the glfs_file_lock() get stucked
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717824
--- Comment #2 from Xiubo Li ---
Created attachment 1577820
--> https://bugzilla.redhat.com/attachment.cgi?id=1577820&action=edit
pstack of on node rhel1
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 09:41:50 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:41:50 +0000
Subject: [Bugs] [Bug 1717827] New: tests/geo-rep: Add test case to validate
non-root geo-replication setup
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717827
Bug ID: 1717827
Summary: tests/geo-rep: Add test case to validate non-root
geo-replication setup
Product: GlusterFS
Version: mainline
Status: NEW
Component: geo-replication
Assignee: bugs at gluster.org
Reporter: sunkumar at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
Add test case to validate non-root geo-replication setup.
Version-Release number of selected component (if applicable):
mainline
How reproducible:
Steps to Reproduce:
1.
2.
3.
Actual results:
Expected results:
Additional info:
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 09:42:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:42:06 +0000
Subject: [Bugs] [Bug 1717827] tests/geo-rep: Add test case to validate
non-root geo-replication setup
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717827
Sunny Kumar changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
Assignee|bugs at gluster.org |sunkumar at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 09:42:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 09:42:10 +0000
Subject: [Bugs] [Bug 1717824] Fencing: Added the tcmu-runner ALUA feature
support but after one of node is rebooted the glfs_file_lock() get stucked
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717824
--- Comment #3 from Xiubo Li ---
The bt output from the gbd:
[root at rhel1 ~]# gdb -p 1325
(gdb) bt
#0 0x00007fc7761baf47 in pthread_join () from /lib64/libpthread.so.0
#1 0x00007fc7773de468 in event_dispatch_epoll (event_pool=0x559f03d4b560) at
event-epoll.c:847
#2 0x0000559f02419658 in main (argc=21, argv=0x7fff9c6722c8) at
glusterfsd.c:2871
(gdb)
[root at rhel3 ~]# gdb -p 7669
(gdb) bt
#0 0x00007fac80bd9f47 in pthread_join () from /usr/lib64/libpthread.so.0
#1 0x00007fac81dfd468 in event_dispatch_epoll (event_pool=0x55de6f845560) at
event-epoll.c:847
#2 0x000055de6f143658 in main (argc=21, argv=0x7ffcafc3eff8) at
glusterfsd.c:2871
(gdb)
The pl_inode->fop_wind_count is:
(gdb) thread 2
[Switching to thread 2 (Thread 0x7fc742184700 (LWP 1829))]
#0 0x00007fc7761bd965 in pthread_cond_wait@@GLIBC_2.3.2 () from
/lib64/libpthread.so.0
(gdb) frame 2
#2 0x00007fc76379c13b in pl_lk (frame=frame at entry=0x7fc750001128,
this=this at entry=0x7fc75c0128f0,
fd=fd at entry=0x7fc73c0977d8, cmd=cmd at entry=6,
flock=flock at entry=0x7fc73c076938,
xdata=xdata at entry=0x7fc73c071828) at posix.c:2637
2637 ret = pl_lock_preempt(pl_inode, reqlock);
(gdb) p pl_inode->fop_wind_count
$1 = -30
(gdb)
The pstack logs please see the attachments
Thanks.
BRs
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 10:19:51 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 10:19:51 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22832
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 10:19:52 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 10:19:52 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #682 from Worker Ant ---
REVIEW: https://review.gluster.org/22832 (glusterd: Fix a typo) posted (#1) for
review on master by Anoop C S
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 11:40:44 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 11:40:44 +0000
Subject: [Bugs] [Bug 1717876] New: Gluster upstream regression tests are
failing with centos 7.7
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717876
Bug ID: 1717876
Summary: Gluster upstream regression tests are failing with
centos 7.7
Product: GlusterFS
Version: mainline
Status: NEW
Component: tests
Assignee: bugs at gluster.org
Reporter: khiremat at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
Gluster upstream regression tests are failing with centos 7.7 with python3
package delivered. The python test utility files are not python3 compatible.
Following are the test utility python files
/bugs/distribute/overlap.py
./bugs/nfs/socket-as-fifo.py
./features/ipctest.py
./utils/create-files.py
./utils/getfattr.py
./utils/gfid-access.py
./utils/libcxattr.py
./utils/pidof.py
./utils/setfattr.py
./utils/changelogparser.py
Each needs to be tested and made py2/py3 compatible
Version-Release number of selected component (if applicable):
mainline
How reproducible:
Always
Failures:
https://build.gluster.org/job/centos7-regression/6317/consoleFull
&https://build.gluster.org/job/centos7-regression/6316/consoleFull
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 11:43:55 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 11:43:55 +0000
Subject: [Bugs] [Bug 1717876] Gluster upstream regression tests are failing
with centos 7.7
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717876
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22833
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 11:43:56 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 11:43:56 +0000
Subject: [Bugs] [Bug 1717876] Gluster upstream regression tests are failing
with centos 7.7
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717876
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22833 (tests: Use python2 for tests) posted
(#1) for review on master by Kotresh HR
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 11:57:16 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 11:57:16 +0000
Subject: [Bugs] [Bug 1708929] Add more test coverage for shd mux
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1708929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-06 11:57:16
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22697 (tests/shd: Add test coverage for shd
mux) merged (#15) on master by Pranith Kumar Karampuri
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 13:29:44 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 13:29:44 +0000
Subject: [Bugs] [Bug 1716760] Make debugging hung frames easier
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716760
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |amukherj at redhat.com
--- Comment #3 from Atin Mukherjee ---
Vivek/Rahul - we need this patch in 3.5.0 for better debugging experience.
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 14:09:30 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 14:09:30 +0000
Subject: [Bugs] [Bug 1717953] New: SELinux context labels are missing for
newly added bricks using add-brick command
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717953
Bug ID: 1717953
Summary: SELinux context labels are missing for newly added
bricks using add-brick command
Product: GlusterFS
Version: mainline
OS: Linux
Status: NEW
Component: scripts
Severity: medium
Assignee: bugs at gluster.org
Reporter: anoopcs at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
When we add new bricks to an existing volume using add-brick command
"glusterd_brick_t" SELinux context label is not assigned on those new brick
paths.
Version-Release number of selected component (if applicable):
master
How reproducible:
Always
Steps to Reproduce:
1. Create and start a basic distribute-replicate volume
2. Verify that brick paths have "glusted_brick_t" SELinux labels by running `ls
-lZ `
3. Add new bricks to the existing volume
4. Check SELinux labels on newly added brick paths
Actual results:
"glusterd_brick_t" SELinux label is missing on newly added bricks
Expected results:
Following SELinux label is expected:
system_u:object_r:glusterd_brick_t:s0
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 14:10:00 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 14:10:00 +0000
Subject: [Bugs] [Bug 1717953] SELinux context labels are missing for newly
added bricks using add-brick command
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717953
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
Assignee|bugs at gluster.org |anoopcs at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Thu Jun 6 14:16:17 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 14:16:17 +0000
Subject: [Bugs] [Bug 1717953] SELinux context labels are missing for newly
added bricks using add-brick command
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717953
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22834
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Thu Jun 6 14:16:18 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Thu, 06 Jun 2019 14:16:18 +0000
Subject: [Bugs] [Bug 1717953] SELinux context labels are missing for newly
added bricks using add-brick command
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717953
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|ASSIGNED |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22834 (extras/hooks: Add SELinux label on
new bricks during add-brick) posted (#1) for review on master by Anoop C S
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 05:02:31 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 05:02:31 +0000
Subject: [Bugs] [Bug 1651445] [RFE] storage.reserve option should take size
of disk as input instead of percentage
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1651445
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1573077
Depends On|1573077 |
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1573077
[Bug 1573077] [RFE] storage.reserve option should take size of disk as input
instead of percentage
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 05:38:38 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 05:38:38 +0000
Subject: [Bugs] [Bug 1715422] ctime: Upgrade/Enabling ctime feature wrongly
updates older files with latest {a|m|c}time
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1715422
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |amukherj at redhat.com
Flags| |needinfo?(khiremat at redhat.c
| |om)
--- Comment #2 from Atin Mukherjee ---
Kotresh - I believe we need to fix this. Can we have a devel ack on this bug?
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 07:00:58 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 07:00:58 +0000
Subject: [Bugs] [Bug 1715422] ctime: Upgrade/Enabling ctime feature wrongly
updates older files with latest {a|m|c}time
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1715422
RHEL Product and Program Management changed:
What |Removed |Added
----------------------------------------------------------------------------
Rule Engine Rule| |Gluster: set
| |qe_test_coverage flag at QE
| |approved BZs
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 08:22:54 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:22:54 +0000
Subject: [Bugs] [Bug 1718191] New: Regression: Intermittent test failure for
quick-read-with-upcall.t
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718191
Bug ID: 1718191
Summary: Regression: Intermittent test failure for
quick-read-with-upcall.t
Product: GlusterFS
Version: mainline
Status: NEW
Component: quick-read
Severity: urgent
Priority: high
Assignee: bugs at gluster.org
Reporter: atumball at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
While running regression, quick-read-with-upcall.t script fails intermittently.
Please debug and fix the problem.
Version-Release number of selected component (if applicable):
master
How reproducible:
1/10
Steps to Reproduce:
1. Submit a patch and run regression.
Additional info:
Error is normally like below:
08:59:24 ok 11 [ 10/ 3] < 36> 'write_to /mnt/glusterfs/0/test.txt
test-message1'
08:59:24 ok 12 [ 10/ 6] < 37> 'test-message1 cat
/mnt/glusterfs/0/test.txt'
08:59:24 ok 13 [ 10/ 4] < 38> 'test-message0 cat
/mnt/glusterfs/1/test.txt'
08:59:24 not ok 14 [ 3715/ 6] < 45> 'test-message1 cat
/mnt/glusterfs/1/test.txt' -> 'Got "test-message0" instead of "test-message1"'
08:59:24 ok 15 [ 10/ 162] < 47> 'gluster --mode=script --wignore
volume set patchy features.cache-invalidation on'
08:59:24 ok 16 [ 10/ 148] < 48> 'gluster --mode=script --wignore
volume set patchy performance.qr-cache-timeout 15'
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:29:07 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:29:07 +0000
Subject: [Bugs] [Bug 1718191] Regression: Intermittent test failure for
quick-read-with-upcall.t
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718191
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22836
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:29:08 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:29:08 +0000
Subject: [Bugs] [Bug 1718191] Regression: Intermittent test failure for
quick-read-with-upcall.t
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718191
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22836 (tests/quick-read-upcall: mark it bad)
posted (#1) for review on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:30:08 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:30:08 +0000
Subject: [Bugs] [Bug 1717757] BItrot: Segmentation Fault if bitrot stub do
signature
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
--- Comment #4 from david.spisla at iternity.com ---
@Amar
I wrote a patch with debug logs and I will observe the bricks now. During this
time I have some questions concerning your patch suggestion:
1. According to crash report from the brick locks, there was a failure in
[2019-06-05 09:05:05.137156] I [MSGID: 139001]
[posix-acl.c:263:posix_acl_log_permit_denied] 0-archive1-access-control:
client: CTX_ID:fcab5e67-b9d9-4b72-8c15-f29de2084af3-GRAPH_ID:0-PID:189
16-HOST:fs-detlefh-c1-n2-PC_NAME:archive1-client-0-RECON_NO:-0, gfid:
494b42ad-7e40-4e27-8878-99387a80b5dc, req(uid:2000,gid:2000,perm:3,ngrps:1),
ctx(uid:0,gid:0,in-groups:0,perm:755,update
d-fop:LOOKUP, acl:-) [Permission denied]
just before the crash. What can be the reason for this?
2. If this LOOKUP for acls fails, is it problematic to do a setxattr with loc?
If we skip setting xattr when fd is NULL the file on that brick won't have the
necessary xattr like trusted.worm_file and other. See an example directly after
the crash:
# file: gluster/brick3/glusterbrick/test/data/BC/storage.log
trusted.gfid=0sag3y6RuoTgqAw//fx3ZB1Q==
trusted.gfid2path.273f2255a25b2961="bd910b86-d51a-4006-a2c4-515ef5f1777a/storage.log"
trusted.pgfid.bd910b86-d51a-4006-a2c4-515ef5f1777a=0sAAAAAQ==
On the healthy brick I got:
# file: gluster/brick3/glusterbrick/test/data/BC/storage.log
trusted.afr.dirty=0sAAAAAAAAAAAAAAAA
trusted.afr.test-client-0=0sAAAABAAAAAMAAAAA
trusted.bit-rot.version=0sAgAAAAAAAABc+P64AAEhGQ==
trusted.gfid=0sag3y6RuoTgqAw//fx3ZB1Q==
trusted.gfid2path.273f2255a25b2961="bd910b86-d51a-4006-a2c4-515ef5f1777a/storage.log"
trusted.glusterfs.mdata=0sAQAAAAAAAAAAAAAAAFz5AJEAAAAAMqdgMwAAAABcRwJEAAAAAAAAAAAAAAAAXPkAkQAAAAAAAAAA
trusted.pgfid.bd910b86-d51a-4006-a2c4-515ef5f1777a=0sAAAAAQ==
trusted.start_time="1559822481"
trusted.worm_file=0sMQA=
After restarting the faulty brick a heal was triggered and afterwards the file
on the faulty brick is heal.It should be ensured that the broken file gets all
necessary xattr. What is the better way? Triggering a setxattr with loc in
worm_create_cbk or do a heal afterwards?
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Fri Jun 7 08:31:55 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:31:55 +0000
Subject: [Bugs] [Bug 1708929] Add more test coverage for shd mux
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1708929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22837
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:31:56 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:31:56 +0000
Subject: [Bugs] [Bug 1708929] Add more test coverage for shd mux
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1708929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|CLOSED |POST
Resolution|NEXTRELEASE |---
Keywords| |Reopened
--- Comment #3 from Worker Ant ---
REVIEW: https://review.gluster.org/22837 (tests/volume-scale-shd-mux: mark as
bad test) posted (#1) for review on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:35:33 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:35:33 +0000
Subject: [Bugs] [Bug 1717757] BItrot: Segmentation Fault if bitrot stub do
signature
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
--- Comment #5 from Amar Tumballi ---
1. permission denied is mostly probably a issue of missing permission (uid
2000, trying to create an entry in a directory with 755, owned by uid-0
(root)).
2. I think it is better to leave it to heal. If it is a create failure, we
should anyways fail the operation is my opinion.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Fri Jun 7 08:40:14 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:40:14 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22835
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:40:15 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:40:15 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #683 from Worker Ant ---
REVIEW: https://review.gluster.org/22835 (tests/subdir-mount: give more time
for heal to complete) posted (#2) for review on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:41:35 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:41:35 +0000
Subject: [Bugs] [Bug 1708929] Add more test coverage for shd mux
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1708929
Amar Tumballi changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |atumball at redhat.com
--- Comment #4 from Amar Tumballi ---
Reopened because of the test script failure:
volume-scale-shd-mux.t
09:09:24 not ok 58 [ 14/ 80343] < 104> '^3$ number_healer_threads_shd
patchy_distribute1 __afr_shd_healer_wait' -> 'Got "1" instead of "^3$"'
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:49:00 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:49:00 +0000
Subject: [Bugs] [Bug 1708929] Add more test coverage for shd mux
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1708929
Amar Tumballi changed:
What |Removed |Added
----------------------------------------------------------------------------
Priority|unspecified |high
Severity|unspecified |high
--- Comment #5 from Worker Ant ---
REVIEW: https://review.gluster.org/22837 (tests/volume-scale-shd-mux: mark as
bad test) merged (#1) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 08:49:43 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 08:49:43 +0000
Subject: [Bugs] [Bug 1718191] Regression: Intermittent test failure for
quick-read-with-upcall.t
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718191
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22836 (tests/quick-read-upcall: mark it bad)
merged (#2) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 09:05:16 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 09:05:16 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #684 from Worker Ant ---
REVIEW: https://review.gluster.org/22829 (tests/utils: Fix py2/py3 util python
scripts) merged (#4) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 09:32:08 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 09:32:08 +0000
Subject: [Bugs] [Bug 1714536] geo-rep: With heavy rename workload geo-rep
log if flooded
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714536
Sunil Kumar Acharya changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |sheggodu at redhat.com
Flags| |needinfo?(khiremat at redhat.c
| |om)
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 09:57:15 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 09:57:15 +0000
Subject: [Bugs] [Bug 1717757] BItrot: Segmentation Fault if bitrot stub do
signature
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717757
--- Comment #6 from david.spisla at iternity.com ---
Allrigth, I will stress the system for a while and if everything is stable I
will commit the patch to gerrit
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the Docs Contact for the bug.
From bugzilla at redhat.com Fri Jun 7 10:34:53 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 10:34:53 +0000
Subject: [Bugs] [Bug 1718227] New: SELinux context labels are missing for
newly added bricks using add-brick command
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718227
Bug ID: 1718227
Summary: SELinux context labels are missing for newly added
bricks using add-brick command
Product: GlusterFS
Version: 6
OS: Linux
Status: NEW
Component: scripts
Severity: medium
Assignee: bugs at gluster.org
Reporter: anoopcs at redhat.com
CC: bugs at gluster.org
Depends On: 1717953
Target Milestone: ---
Classification: Community
+++ This bug was initially created as a clone of Bug #1717953 +++
Description of problem:
When we add new bricks to an existing volume using add-brick command
"glusterd_brick_t" SELinux context label is not assigned on those new brick
paths.
Version-Release number of selected component (if applicable):
master
How reproducible:
Always
Steps to Reproduce:
1. Create and start a basic distribute-replicate volume
2. Verify that brick paths have "glusted_brick_t" SELinux labels by running `ls
-lZ `
3. Add new bricks to the existing volume
4. Check SELinux labels on newly added brick paths
Actual results:
"glusterd_brick_t" SELinux label is missing on newly added bricks
Expected results:
Following SELinux label is expected:
system_u:object_r:glusterd_brick_t:s0
--- Additional comment from Worker Ant on 2019-06-06 19:46:18 IST ---
REVIEW: https://review.gluster.org/22834 (extras/hooks: Add SELinux label on
new bricks during add-brick) posted (#1) for review on master by Anoop C S
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1717953
[Bug 1717953] SELinux context labels are missing for newly added bricks using
add-brick command
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 10:34:53 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 10:34:53 +0000
Subject: [Bugs] [Bug 1717953] SELinux context labels are missing for newly
added bricks using add-brick command
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717953
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1718227
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1718227
[Bug 1718227] SELinux context labels are missing for newly added bricks using
add-brick command
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 10:36:15 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 10:36:15 +0000
Subject: [Bugs] [Bug 1718227] SELinux context labels are missing for newly
added bricks using add-brick command
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718227
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1686800
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 10:55:40 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 10:55:40 +0000
Subject: [Bugs] [Bug 1678640] Running 'control-cpu-load.sh' prevents CTDB
starting
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1678640
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |anoopcs at redhat.com
Flags| |needinfo?(ryan at magenta.tv)
--- Comment #1 from Anoop C S ---
(In reply to ryan from comment #0)
> Actual results:
> CTDB fails to start with following error:
> 2019/02/08 20:46:59.612215 ctdbd[2629]: Created PID file
> /var/run/ctdb/ctdbd.pid
> 2019/02/08 20:46:59.612267 ctdbd[2629]: Listening to ctdb socket
> /var/run/ctdb/ctdbd.socket
> 2019/02/08 20:46:59.612297 ctdbd[2629]: Unable to set scheduler to
> SCHED_FIFO (Operation not permitted)
> 2019/02/08 20:46:59.612304 ctdbd[2629]: CTDB daemon shutting down
Please use the following CTDB setting in /etc/sysconfig/ctdb:
CTDB_NOSETSCHED=yes
and try restarting CTDB.
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:01:49 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:01:49 +0000
Subject: [Bugs] [Bug 1678640] Running 'control-cpu-load.sh' prevents CTDB
starting
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1678640
--- Comment #2 from Anoop C S ---
(In reply to Anoop C S from comment #1)
> (In reply to ryan from comment #0)
> > Actual results:
> > CTDB fails to start with following error:
> > 2019/02/08 20:46:59.612215 ctdbd[2629]: Created PID file
> > /var/run/ctdb/ctdbd.pid
> > 2019/02/08 20:46:59.612267 ctdbd[2629]: Listening to ctdb socket
> > /var/run/ctdb/ctdbd.socket
> > 2019/02/08 20:46:59.612297 ctdbd[2629]: Unable to set scheduler to
> > SCHED_FIFO (Operation not permitted)
> > 2019/02/08 20:46:59.612304 ctdbd[2629]: CTDB daemon shutting down
>
> Please use the following CTDB setting in /etc/sysconfig/ctdb:
> CTDB_NOSETSCHED=yes
>
> and try restarting CTDB.
Copy-pasting a summary of the reason for above suggestion from a different bug:
CTDB daemon i.e, ctdbd is a service that by default requests for real-time
scheduling unless it is instructed not to do so via explicit configuration
parameters. By default systemd places all system services into their own
control groups in the "cpu" hierarchy. But the "cpu" cgroup controller of the
kernel demands absolute real-time budget to be explicitly specified. A
reasonable value for required real-time cpu cycles are pre-written into
corresponding configuration files. This value getting overwritten by other
components in the system results in denial of real-time scheduling to services
under this "cpu" hierarchy with error EPERM(Operation not permitted).
ref:
https://www.freedesktop.org/wiki/Software/systemd/MyServiceCantGetRealtime/
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:04:59 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:04:59 +0000
Subject: [Bugs] [Bug 1678640] Running 'control-cpu-load.sh' prevents CTDB
starting
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1678640
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |ASSIGNED
Component|core |gluster-smb
Assignee|moagrawa at redhat.com |anoopcs at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:16:19 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:16:19 +0000
Subject: [Bugs] [Bug 1680085] OS X clients disconnect from SMB mount points
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1680085
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
Version|cns-1.0 |4.1
Component|samba |gluster-smb
CC| |bugs at gluster.org
Assignee|gdeschner at redhat.com |bugs at gluster.org
QA Contact|vdas at redhat.com |
Product|Red Hat Gluster Storage |GlusterFS
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 11:16:46 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:16:46 +0000
Subject: [Bugs] [Bug 1714536] geo-rep: With heavy rename workload geo-rep
log if flooded
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714536
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Flags|needinfo?(khiremat at redhat.c |
|om) |
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:19:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:19:10 +0000
Subject: [Bugs] [Bug 1602824] SMBD crashes when streams_attr VFS is used
with Gluster VFS
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1602824
Anoop C S changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |anoopcs at redhat.com
Component|libgfapi |gluster-smb
QA Contact|bugs at gluster.org |
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 11:20:40 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:20:40 +0000
Subject: [Bugs] [Bug 1717876] Gluster upstream regression tests are failing
with centos 7.7
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717876
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |MODIFIED
--- Comment #2 from Kotresh HR ---
This patch fixed the issue and is merged
https://review.gluster.org/#/c/glusterfs/+/22829/
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 11:21:07 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:21:07 +0000
Subject: [Bugs] [Bug 1717876] Gluster upstream regression tests are failing
with centos 7.7
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717876
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Assignee|bugs at gluster.org |khiremat at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 11:21:29 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:21:29 +0000
Subject: [Bugs] [Bug 1717876] Gluster upstream regression tests are failing
with centos 7.7
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717876
Kotresh HR changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|MODIFIED |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-07 11:21:29
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:26:59 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:26:59 +0000
Subject: [Bugs] [Bug 1714536] geo-rep: With heavy rename workload geo-rep
log if flooded
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1714536
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |MODIFIED
CC| |amukherj at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:49:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:49:10 +0000
Subject: [Bugs] [Bug 1718273] New: markdown formatting errors in files
present under /doc directory of the project
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718273
Bug ID: 1718273
Summary: markdown formatting errors in files present under /doc
directory of the project
Product: GlusterFS
Version: mainline
Status: NEW
Component: doc
Keywords: Documentation
Assignee: kiyer at redhat.com
Reporter: kiyer at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
There are a lot of markdown files present under /doc directory of the project
which are having markdown formatting errors which make these files look really
shabby when open on GitHub.
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:53:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:53:06 +0000
Subject: [Bugs] [Bug 1718273] markdown formatting errors in files present
under /doc directory of the project
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718273
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22825
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 11:53:07 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 11:53:07 +0000
Subject: [Bugs] [Bug 1718273] markdown formatting errors in files present
under /doc directory of the project
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718273
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22825 (Fixing formatting errors in markdown
files) posted (#2) for review on master by Kshithij Iyer
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Fri Jun 7 13:11:43 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 13:11:43 +0000
Subject: [Bugs] [Bug 1718316] New: Ganesha-gfapi logs are flooded with error
messages related to "gf_uuid_is_null(gfid)) [Invalid argument]" when
lookups are running from multiple clients
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718316
Bug ID: 1718316
Summary: Ganesha-gfapi logs are flooded with error messages
related to "gf_uuid_is_null(gfid)) [Invalid argument]"
when lookups are running from multiple clients
Product: GlusterFS
Version: mainline
Hardware: All
OS: All
Status: NEW
Component: libgfapi
Severity: high
Priority: medium
Assignee: bugs at gluster.org
Reporter: skoduri at redhat.com
QA Contact: bugs at gluster.org
CC: bugs at gluster.org, dang at redhat.com, ffilz at redhat.com,
grajoria at redhat.com, jthottan at redhat.com,
mbenjamin at redhat.com, msaini at redhat.com,
rhs-bugs at redhat.com, sankarshan at redhat.com,
skoduri at redhat.com, storage-qa-internal at redhat.com
Depends On: 1717784
Target Milestone: ---
Classification: Community
+++ This bug was initially created as a clone of Bug #1717784 +++
Description of problem:
=========================
Ganesha-gfapi logs are flooded with errors messages related to
gf_uuid_is_null(gfid), when linux untars and lookups are running from multiple
clients-
---------
[2019-06-06 07:56:12.503603] E [glfs-handleops.c:1892:glfs_h_find_handle]
(-->/lib64/libgfapi.so.0(+0xe0ae) [0x7f7e91e8b0ae]
-->/lib64/libgfapi.so.0(+0x258f1) [0x7f7e91ea28f1]
-->/lib64/libgfapi.so.0(+0x257c4) [0x7f7e91ea27c4] ) 0-glfs_h_find_handle:
invalid argument: !(gf_uuid_is_null(gfid)) [Invalid argument]
---------
Version-Release number of selected component (if applicable):
===========================
# cat /etc/redhat-release
Red Hat Enterprise Linux Server release 7.7 Beta (Maipo)
# rpm -qa | grep ganesha
nfs-ganesha-2.7.3-3.el7rhgs.x86_64
glusterfs-ganesha-6.0-3.el7rhgs.x86_64
nfs-ganesha-debuginfo-2.7.3-3.el7rhgs.x86_64
nfs-ganesha-gluster-2.7.3-3.el7rhgs.x86_64
How reproducible:
=====================
2/2
Steps to Reproduce:
======================
1.Create 4 node Ganesha cluster
2.Create 4*3 Distribute-replicate Volume.Export the volume via Ganesha
3.Mount the volume on 4 clients via v4.1 protocol
4.Run the following workload
Client 1: Run linux untars
Client 2: du -sh in loop
Client 3: ls -lRt in loop
Client 4: find's in loop
Actual results:
==================
While test is running,ganesha-gfapi logs are flooded with errors related to
"gf_uuid_is_null"
======
[2019-06-03 16:54:19.829136] E [glfs-handleops.c:1892:glfs_h_find_handle]
(-->/lib64/libgfapi.so.0(+0xe0ae) [0x7ff6902d00ae]
-->/lib64/libgfapi.so.0(+0x2594a) [0x7ff6902e794a]
-->/lib64/libgfapi.so.0(+0x257c4) [0x7ff6902e77c4] ) 0-glfs_h_find_handle:
invalid argument: !(gf_uuid_is_null(gfid)) [Invalid argument]
[2019-06-03 16:54:20.006163] E [glfs-handleops.c:1892:glfs_h_find_handle]
(-->/lib64/libgfapi.so.0(+0xe0ae) [0x7ff6902d00ae]
-->/lib64/libgfapi.so.0(+0x2594a) [0x7ff6902e794a]
-->/lib64/libgfapi.so.0(+0x257c4) [0x7ff6902e77c4] ) 0-glfs_h_find_handle:
invalid argument: !(gf_uuid_is_null(gfid)) [Invalid argument]
[2019-06-03 16:54:20.320293] E [glfs-handleops.c:1892:glfs_h_find_handle]
(-->/lib64/libgfapi.so.0(+0xe0ae) [0x7ff6902d00ae]
-->/lib64/libgfapi.so.0(+0x2594a) [0x7ff6902e794a]
-->/lib64/libgfapi.so.0(+0x257c4) [0x7ff6902e77c4] ) 0-glfs_h_find_handle:
invalid argument: !(gf_uuid_is_null(gfid)) [Invalid argument]
=====
# cat /var/log/ganesha/ganesha-gfapi.log | grep gf_uuid_is_null | wc -l
605340
Expected results:
===================
There should not be error messages in ganesha-gfapi.logs
Additional info:
===================
On narrowing down the test scenario,Seems to be the error messages are coming
when only du -sh and ls -lRt are running in loop from two different clients
--- Additional comment from RHEL Product and Program Management on 2019-06-06
08:10:27 UTC ---
This bug is automatically being proposed for the next minor release of Red Hat
Gluster Storage by setting the release flag 'rhgs?3.5.0' to '?'.
If this bug should be proposed for a different release, please manually change
the proposed release flag.
--- Additional comment from Soumya Koduri on 2019-06-06 09:48:36 UTC ---
@Manisha,
are these clients connected to different NFS-Ganesha servers? On which machine
did you observe these errors? I do not see such messages in the sosreports
uploaded.
>>> On narrowing down the test scenario,Seems to be the error messages are coming when only du -sh and ls -lRt are running in loop from two different clients
Does this mean, these messages are not seen with just linux untar test?
--- Additional comment from Manisha Saini on 2019-06-06 10:16:00 UTC ---
(In reply to Soumya Koduri from comment #3)
> @Manisha,
>
> are these clients connected to different NFS-Ganesha servers? On which
> machine did you observe these errors? I do not see such messages in the
> sosreports uploaded.
Hi soumya,
All the clients are connected to single server VIP
I see there is some issue with how sosreport collecting ganesha logs.All logs
are not captured as part of sosreport.
>
> >>> On narrowing down the test scenario,Seems to be the error messages are coming when only du -sh and ls -lRt are running in loop from two different clients
>
> Does this mean, these messages are not seen with just linux untar test?
No.Not seen with only untars
--- Additional comment from Soumya Koduri on 2019-06-07 10:08:03 UTC ---
Thanks Manisha for sharing the setup and logs.
"0-glfs_h_find_handle: invalid argument: !(gf_uuid_is_null(gfid)) [Invalid
argument] "
The above message is logged while processing upcall requests. Somehow the gfid
passed has become NULL. IMO there are two issues to be considered here -
> there are so many upcall requests generated even though there is only single server serving all the clients.
Seems like the data being accessed is huge and hence the server is trying clean
up the inodes from the lru list. While destroying a inode, upcall xlator sends
cache invalidation request to all its clients to notify that the particular
file/inode entry is no more cached by the server. This logic can be optimized a
bit here.
For nameless lookups, server generates a dummy inode (say inodeD) and later
links it to inode (if there is no entry already present for that file/dir) in
the cbk path. So as part of lookup_cbk, though the inode (inodeD) received is
invalid, upcall xlator creates an inode_ctx entry as it eventually can get
linked to the inode table. However in certain cases, if there is already an
inode (say inodeC) present for that particular file, this new inode (inodeD)
created will be purged, which results in sending upcall notifications to the
clients.
in Manisha's testcase, as the data created is huge and being looked up in a
loop, there are many such dummy inode entries getting purged resulting in huge
number of upcall notifications sent to the client. We can avoid this issue to
an extent by checking if the inode is valid or not (i.e, linked or not) before
sending callback notifications.
note - this has been day-1 issue but good to be fixed.
* Another issue is gfid becoming NULL in upcall args.
> I couldn't reproduce this issue on my setup. However seems like in upcall xlator we already check if the gfid is not NULL before sending notification.
GF_VALIDATE_OR_GOTO("upcall_client_cache_invalidate",
!(gf_uuid_is_null(gfid)), out);
So that means somewhere in the client processing, gfid has become NULL. From
further code-reading I see a potential issue in upcall processing callback
function -
In glfs_cbk_upcall_data(),
--
args->fs = fs;
args->upcall_data = gf_memdup(upcall_data, sizeof(*upcall_data));
--
gf_memdup() may not be the right routine to use here as upcall_data structure
contains pointers to other data. This definitely needs to be fixed. However
would like to re-confirm if this caused gfid to become NULL. Request Manisha to
share setup (if possible) while the tests going on to confirm this theory.
Thanks!
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1717784
[Bug 1717784] Ganesha-gfapi logs are flooded with error messages related to
"gf_uuid_is_null(gfid)) [Invalid argument]" when lookups are running from
multiple clients
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 13:14:59 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 13:14:59 +0000
Subject: [Bugs] [Bug 1193929] GlusterFS can be improved
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1193929
--- Comment #685 from Worker Ant ---
REVIEW: https://review.gluster.org/22832 (glusterd: Fix typos) merged (#3) on
master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 13:59:52 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 13:59:52 +0000
Subject: [Bugs] [Bug 1718338] New: Upcall: Avoid sending upcalls for invalid
Inode
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718338
Bug ID: 1718338
Summary: Upcall: Avoid sending upcalls for invalid Inode
Product: GlusterFS
Version: mainline
Hardware: All
OS: All
Status: NEW
Component: upcall
Severity: high
Assignee: bugs at gluster.org
Reporter: skoduri at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
For nameless LOOKUPs, server creates a new inode which shall remain invalid
until the fop is successfully processed post which it is linked to the inode
table.
But incase if there is an already linked inode for that entry, it discards that
newly created inode which results in upcall notification. This may result in
client being bombarded with unnecessary upcalls affecting performance if the
data set is huge.
This issue can be avoided by looking up and storing the upcall context in the
original linked inode (if exists), thus saving up on those extra callbacks.
Version-Release number of selected component (if applicable):
How reproducible:
Steps to Reproduce:
1.
2.
3.
Actual results:
Expected results:
Additional info:
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 14:08:38 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 14:08:38 +0000
Subject: [Bugs] [Bug 1718316] Ganesha-gfapi logs are flooded with error
messages related to "gf_uuid_is_null(gfid)) [Invalid argument]" when
lookups are running from multiple clients
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718316
Soumya Koduri changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1717784
Depends On|1717784 |
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1717784
[Bug 1717784] Ganesha-gfapi logs are flooded with error messages related to
"gf_uuid_is_null(gfid)) [Invalid argument]" when lookups are running from
multiple clients
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 14:08:38 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 14:08:38 +0000
Subject: [Bugs] [Bug 1718338] Upcall: Avoid sending upcalls for invalid Inode
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718338
Soumya Koduri changed:
What |Removed |Added
----------------------------------------------------------------------------
Blocks| |1717784
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1717784
[Bug 1717784] Ganesha-gfapi logs are flooded with error messages related to
"gf_uuid_is_null(gfid)) [Invalid argument]" when lookups are running from
multiple clients
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 14:09:42 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 14:09:42 +0000
Subject: [Bugs] [Bug 1718316] Ganesha-gfapi logs are flooded with error
messages related to "gf_uuid_is_null(gfid)) [Invalid argument]" when
lookups are running from multiple clients
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718316
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22839
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 14:09:44 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 14:09:44 +0000
Subject: [Bugs] [Bug 1718316] Ganesha-gfapi logs are flooded with error
messages related to "gf_uuid_is_null(gfid)) [Invalid argument]" when
lookups are running from multiple clients
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718316
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22839 (gfapi: fix incorrect initialization
of upcall syncop arguments) posted (#1) for review on master by soumya k
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 14:10:25 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 14:10:25 +0000
Subject: [Bugs] [Bug 1718316] Ganesha-gfapi logs are flooded with error
messages related to "gf_uuid_is_null(gfid)) [Invalid argument]" when
lookups are running from multiple clients
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718316
Soumya Koduri changed:
What |Removed |Added
----------------------------------------------------------------------------
Keywords| |Triaged
Assignee|bugs at gluster.org |skoduri at redhat.com
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 14:10:52 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 14:10:52 +0000
Subject: [Bugs] [Bug 1718338] Upcall: Avoid sending upcalls for invalid Inode
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718338
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22840
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Fri Jun 7 14:11:30 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Fri, 07 Jun 2019 14:11:30 +0000
Subject: [Bugs] [Bug 1718338] Upcall: Avoid sending upcalls for invalid Inode
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718338
Soumya Koduri changed:
What |Removed |Added
----------------------------------------------------------------------------
Keywords| |Triaged
Assignee|bugs at gluster.org |skoduri at redhat.com
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 02:13:07 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 02:13:07 +0000
Subject: [Bugs] [Bug 1717782] gluster v get all still showing
storage.fips-mode-rchecksum off
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1717782
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-08 02:13:07
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22830 (glusterd: store fips-mode-rchecksum
option in the info file) merged (#3) on master by Atin Mukherjee
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sat Jun 8 05:40:05 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 05:40:05 +0000
Subject: [Bugs] [Bug 1716766] [Thin-arbiter] TA process is not picking 24007
as port while starting up
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1716766
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-08 05:40:05
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22811 (cluster/replicate: Modify command in
unit file to assign port correctly) merged (#4) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sat Jun 8 05:42:56 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 05:42:56 +0000
Subject: [Bugs] [Bug 1715921] uss.t tests times out with brick-mux regression
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1715921
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-08 05:42:56
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22728 (uss: Ensure that snapshot is deleted
before creating a new snapshot) merged (#10) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 05:46:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 05:46:10 +0000
Subject: [Bugs] [Bug 1718273] markdown formatting errors in files present
under /doc directory of the project
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718273
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed| |2019-06-08 05:46:10
--- Comment #2 from Worker Ant ---
REVIEW: https://review.gluster.org/22825 (Fixing formatting errors in markdown
files) merged (#3) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sat Jun 8 05:47:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 05:47:06 +0000
Subject: [Bugs] [Bug 1703948] Self-heal daemon resources are not cleaned
properly after a ec fini
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703948
--- Comment #7 from Worker Ant ---
REVIEW: https://review.gluster.org/22810 (xlator/log: Add more logging in
xlator_is_cleanup_starting) merged (#4) on master by Amar Tumballi
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sat Jun 8 08:14:37 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 08:14:37 +0000
Subject: [Bugs] [Bug 1688226] Brick Still Died After Restart Glusterd &
Glusterfsd Services
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1688226
Eng Khalid Jamal changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |CLOSED
Resolution|--- |NOTABUG
Last Closed| |2019-06-08 08:14:37
--- Comment #4 from Eng Khalid Jamal ---
i think there is no one can solve this issue for me , when i check my brick i
find my disk is completely offline i replace my disk , and i make gluster
replace brick then re balance my volume then heal it every thing is going
write but in future is there any solution for this issue .
Best regards
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 14:40:16 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 14:40:16 +0000
Subject: [Bugs] [Bug 1697986] GlusterFS 5.7 tracker
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1697986
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22842
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 14:40:17 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 14:40:17 +0000
Subject: [Bugs] [Bug 1697986] GlusterFS 5.7 tracker
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1697986
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22842 (doc: Added release notes for 5.7)
posted (#1) for review on release-5 by hari gowtham
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 14:41:50 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 14:41:50 +0000
Subject: [Bugs] [Bug 1718555] New: (glusterfs-6.3) - GlusterFS 6.3 tracker
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718555
Bug ID: 1718555
Summary: (glusterfs-6.3) - GlusterFS 6.3 tracker
Product: GlusterFS
Version: 4.1
Status: NEW
Component: core
Assignee: bugs at gluster.org
Reporter: hgowtham at redhat.com
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Tracker bug for 6.3
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 15:00:25 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 15:00:25 +0000
Subject: [Bugs] [Bug 1718555] (glusterfs-6.3) - GlusterFS 6.3 tracker
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718555
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
External Bug ID| |Gluster.org Gerrit 22843
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 15:00:26 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 15:00:26 +0000
Subject: [Bugs] [Bug 1718555] (glusterfs-6.3) - GlusterFS 6.3 tracker
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718555
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |POST
--- Comment #1 from Worker Ant ---
REVIEW: https://review.gluster.org/22843 (doc: Added release notes for 6.3)
posted (#1) for review on release-6 by hari gowtham
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 16:34:10 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 16:34:10 +0000
Subject: [Bugs] [Bug 1718562] New: flock failure (regression)
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1718562
Bug ID: 1718562
Summary: flock failure (regression)
Product: GlusterFS
Version: 6
Hardware: x86_64
OS: Linux
Status: NEW
Component: locks
Severity: urgent
Assignee: bugs at gluster.org
Reporter: jaco at uls.co.za
CC: bugs at gluster.org
Target Milestone: ---
Classification: Community
Description of problem:
after a small number of flock rounds the lock remains behind indefinitely until
cleared with volume clear-locks, whereafter which normal operation resumes
again.
I suspect this happens when there is contention on the lock.
I've got a setup where these locks are used syncronization mechanism. So a
process on host a will take the lock, and release it on shutdown, at which
point another host is likely already trying to obtain the lock, and never
manages to do so (clearing granted allows the lock to proceed, but randomly
clearing locks is a high-risk operation).
Version-Release number of selected component (if applicable): glusterfs 6.1
(confirmed working correctly on 3.12.3 and 4.0.2, suspected correct on 4.1.5
but no longer have a setup with 4.1.5 around).
How reproducible: Trivial. In the mentioned application it's on almost every
single lock attempt as far as I can determine.
Steps to Reproduce:
morpheus ~ # gluster volume info shared
Volume Name: shared
Type: Replicate
Volume ID: a4410662-b6e0-4ed0-b1e0-a1cbf168029c
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: morpheus:/mnt/gluster/shared
Brick2: r2d2:/mnt/gluster/shared
Options Reconfigured:
transport.address-family: inet
nfs.disable: on
morpheus ~ # mkdir /mnt/t
morpheus ~ # mount -t glusterfs localhost:shared /mnt/t
morpheus ~ #
r2d2 ~ # mkdir /mnt/t
r2d2 ~ # mount -t glusterfs localhost:shared /mnt/t
r2d2 ~ #
morpheus ~ # cd /mnt/t/
morpheus ~ # ls -l
total 0
morpheus /mnt/t # exec 3>lockfile; c=0; while flock -w 10 -x 3; do (( c++ ));
echo "Iteration $c passed"; exec 3<&-; exec 3>lockfile; done; echo "Failed
after $c iterations"; exec 3<&-
Iteration 1 passed
Iteration 2 passed
Iteration 3 passed
...
r2d2 /mnt/t # exec 3>lockfile; c=0; while flock -w 10 -x 3; do (( c++ )); echo
"Iteration $c passed"; exec 3<&-; exec 3>lockfile; done; echo "Failed after $c
iterations"; exec 3<&-
Iteration 1 passed
Iteration 2 passed
Failed after 2 iterations
r2d2 /mnt/t #
Iteration 100 passed
Iteration 101 passed
Iteration 102 passed
Failed after 102 iterations
morpheus /mnt/t #
The two mounts failed at the same time, morpheus just passed more iterations
due to being started first.
Only iterating on one host I've had to stop it with ^C around 10k iterations,
which to me is sufficient indication that it's contention related.
After the above failure, I need to either rm the file and then it works again,
or I need to issue "gluster volume clear-locks shared /lockfile kind granted
posix"
On /tmp on my local machine I can run as much invocations of the loop above as
I want without issues (ext4 filesystem).
On glusterfs 3.12.3 and 4.0.2 I tried the above too, and stopped them after 10k
iterations.
I have not observed the behaviour on glusterfs 4.1.5 which we used for a very
long time.
I either need a fix for this, or a way (prefereably with little no downtime,
total around 1.8TB of data) to downgrade glusterfs back to 4.1.X. Or a way to
get around this reliably from within my application code (mostly control
scripts written in bash).
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sat Jun 8 05:47:06 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sat, 08 Jun 2019 05:47:06 +0000
Subject: [Bugs] [Bug 1703948] Self-heal daemon resources are not cleaned
properly after a ec fini
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703948
Worker Ant changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
Resolution|--- |NEXTRELEASE
Last Closed|2019-05-22 13:11:42 |2019-06-09 00:31:15
--- Comment #8 from Worker Ant ---
REVIEW: https://review.gluster.org/22798 (ec/fini: Fix race between xlator
cleanup and on going async fop) merged (#12) on master by Pranith Kumar
Karampuri
--
You are receiving this mail because:
You are on the CC list for the bug.
From bugzilla at redhat.com Sun Jun 9 05:34:35 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sun, 09 Jun 2019 05:34:35 +0000
Subject: [Bugs] [Bug 1703007] The telnet or something would cause high
memory usage for glusterd & glusterfsd
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1703007
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|NEW |CLOSED
Resolution|--- |INSUFFICIENT_DATA
Last Closed| |2019-06-09 05:34:35
--- Comment #3 from Atin Mukherjee ---
Closing this as we haven't received sufficient information.
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
From bugzilla at redhat.com Sun Jun 9 05:36:26 2019
From: bugzilla at redhat.com (bugzilla at redhat.com)
Date: Sun, 09 Jun 2019 05:36:26 +0000
Subject: [Bugs] [Bug 1658733]
tests/bugs/glusterd/rebalance-operations-in-single-node.t is failing in
brick mux regression
In-Reply-To:
References:
Message-ID:
https://bugzilla.redhat.com/show_bug.cgi?id=1658733
Atin Mukherjee changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|POST |CLOSED
CC| |amukherj at redhat.com
Resolution|--- |WORKSFORME
Last Closed| |2019-06-09 05:36:26
--- Comment #5 from Atin Mukherjee